Dec 01 17:25:26 crc systemd[1]: Starting Kubernetes Kubelet... Dec 01 17:25:26 crc restorecon[4706]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:26 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 17:25:27 crc restorecon[4706]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 17:25:27 crc restorecon[4706]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 01 17:25:27 crc kubenswrapper[4868]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 17:25:27 crc kubenswrapper[4868]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 01 17:25:27 crc kubenswrapper[4868]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 17:25:27 crc kubenswrapper[4868]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 17:25:27 crc kubenswrapper[4868]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 01 17:25:27 crc kubenswrapper[4868]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.918494 4868 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.921953 4868 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.921975 4868 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.921981 4868 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.921987 4868 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.921996 4868 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922003 4868 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922008 4868 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922012 4868 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922017 4868 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922021 4868 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922025 4868 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922030 4868 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922035 4868 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922040 4868 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922043 4868 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922047 4868 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922051 4868 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922055 4868 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922059 4868 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922063 4868 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922066 4868 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922070 4868 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922074 4868 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922077 4868 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922081 4868 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922085 4868 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922088 4868 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922093 4868 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922098 4868 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922103 4868 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922108 4868 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922114 4868 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922119 4868 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922125 4868 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922131 4868 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922136 4868 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922139 4868 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922144 4868 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922148 4868 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922152 4868 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922156 4868 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922159 4868 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922163 4868 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922167 4868 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922171 4868 feature_gate.go:330] unrecognized feature gate: Example Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922174 4868 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922178 4868 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922181 4868 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922185 4868 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922189 4868 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922194 4868 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922198 4868 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922202 4868 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922206 4868 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922211 4868 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922214 4868 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922220 4868 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922224 4868 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922227 4868 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922232 4868 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922236 4868 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922239 4868 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922243 4868 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922246 4868 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922250 4868 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922253 4868 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922257 4868 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922260 4868 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922264 4868 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922268 4868 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.922272 4868 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922482 4868 flags.go:64] FLAG: --address="0.0.0.0" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922494 4868 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922504 4868 flags.go:64] FLAG: --anonymous-auth="true" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922510 4868 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922517 4868 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922522 4868 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922529 4868 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922536 4868 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922542 4868 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922547 4868 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922552 4868 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922557 4868 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922563 4868 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922568 4868 flags.go:64] FLAG: --cgroup-root="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922572 4868 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922576 4868 flags.go:64] FLAG: --client-ca-file="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922580 4868 flags.go:64] FLAG: --cloud-config="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922585 4868 flags.go:64] FLAG: --cloud-provider="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922589 4868 flags.go:64] FLAG: --cluster-dns="[]" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922594 4868 flags.go:64] FLAG: --cluster-domain="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922598 4868 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922602 4868 flags.go:64] FLAG: --config-dir="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922606 4868 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922611 4868 flags.go:64] FLAG: --container-log-max-files="5" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922617 4868 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922621 4868 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922625 4868 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922630 4868 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922634 4868 flags.go:64] FLAG: --contention-profiling="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922638 4868 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922642 4868 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922646 4868 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922650 4868 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922655 4868 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922660 4868 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922664 4868 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922668 4868 flags.go:64] FLAG: --enable-load-reader="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922673 4868 flags.go:64] FLAG: --enable-server="true" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922677 4868 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922682 4868 flags.go:64] FLAG: --event-burst="100" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922686 4868 flags.go:64] FLAG: --event-qps="50" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922690 4868 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922695 4868 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922699 4868 flags.go:64] FLAG: --eviction-hard="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922704 4868 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922709 4868 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922713 4868 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922717 4868 flags.go:64] FLAG: --eviction-soft="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922721 4868 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922726 4868 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922730 4868 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922734 4868 flags.go:64] FLAG: --experimental-mounter-path="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922738 4868 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922742 4868 flags.go:64] FLAG: --fail-swap-on="true" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922746 4868 flags.go:64] FLAG: --feature-gates="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922751 4868 flags.go:64] FLAG: --file-check-frequency="20s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922756 4868 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922760 4868 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922764 4868 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922769 4868 flags.go:64] FLAG: --healthz-port="10248" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922773 4868 flags.go:64] FLAG: --help="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922777 4868 flags.go:64] FLAG: --hostname-override="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922781 4868 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922785 4868 flags.go:64] FLAG: --http-check-frequency="20s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922789 4868 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922793 4868 flags.go:64] FLAG: --image-credential-provider-config="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922797 4868 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922801 4868 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922805 4868 flags.go:64] FLAG: --image-service-endpoint="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922809 4868 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922814 4868 flags.go:64] FLAG: --kube-api-burst="100" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922819 4868 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922823 4868 flags.go:64] FLAG: --kube-api-qps="50" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922827 4868 flags.go:64] FLAG: --kube-reserved="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922832 4868 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922850 4868 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922855 4868 flags.go:64] FLAG: --kubelet-cgroups="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922859 4868 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922863 4868 flags.go:64] FLAG: --lock-file="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922867 4868 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922872 4868 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922877 4868 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922883 4868 flags.go:64] FLAG: --log-json-split-stream="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922888 4868 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922892 4868 flags.go:64] FLAG: --log-text-split-stream="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922896 4868 flags.go:64] FLAG: --logging-format="text" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922900 4868 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922904 4868 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922908 4868 flags.go:64] FLAG: --manifest-url="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922913 4868 flags.go:64] FLAG: --manifest-url-header="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922919 4868 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922923 4868 flags.go:64] FLAG: --max-open-files="1000000" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922929 4868 flags.go:64] FLAG: --max-pods="110" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922933 4868 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922953 4868 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922958 4868 flags.go:64] FLAG: --memory-manager-policy="None" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922962 4868 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922967 4868 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922973 4868 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922978 4868 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922992 4868 flags.go:64] FLAG: --node-status-max-images="50" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.922998 4868 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923003 4868 flags.go:64] FLAG: --oom-score-adj="-999" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923008 4868 flags.go:64] FLAG: --pod-cidr="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923012 4868 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923020 4868 flags.go:64] FLAG: --pod-manifest-path="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923025 4868 flags.go:64] FLAG: --pod-max-pids="-1" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923029 4868 flags.go:64] FLAG: --pods-per-core="0" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923033 4868 flags.go:64] FLAG: --port="10250" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923037 4868 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923041 4868 flags.go:64] FLAG: --provider-id="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923045 4868 flags.go:64] FLAG: --qos-reserved="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923050 4868 flags.go:64] FLAG: --read-only-port="10255" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923054 4868 flags.go:64] FLAG: --register-node="true" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923061 4868 flags.go:64] FLAG: --register-schedulable="true" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923065 4868 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923073 4868 flags.go:64] FLAG: --registry-burst="10" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923077 4868 flags.go:64] FLAG: --registry-qps="5" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923081 4868 flags.go:64] FLAG: --reserved-cpus="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923085 4868 flags.go:64] FLAG: --reserved-memory="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923092 4868 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923098 4868 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923104 4868 flags.go:64] FLAG: --rotate-certificates="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923109 4868 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923114 4868 flags.go:64] FLAG: --runonce="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923119 4868 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923124 4868 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923129 4868 flags.go:64] FLAG: --seccomp-default="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923134 4868 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923139 4868 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923145 4868 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923149 4868 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923153 4868 flags.go:64] FLAG: --storage-driver-password="root" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923158 4868 flags.go:64] FLAG: --storage-driver-secure="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923162 4868 flags.go:64] FLAG: --storage-driver-table="stats" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923166 4868 flags.go:64] FLAG: --storage-driver-user="root" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923170 4868 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923174 4868 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923179 4868 flags.go:64] FLAG: --system-cgroups="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923183 4868 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923189 4868 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923193 4868 flags.go:64] FLAG: --tls-cert-file="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923198 4868 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923203 4868 flags.go:64] FLAG: --tls-min-version="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923207 4868 flags.go:64] FLAG: --tls-private-key-file="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923211 4868 flags.go:64] FLAG: --topology-manager-policy="none" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923216 4868 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923220 4868 flags.go:64] FLAG: --topology-manager-scope="container" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923225 4868 flags.go:64] FLAG: --v="2" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923232 4868 flags.go:64] FLAG: --version="false" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923238 4868 flags.go:64] FLAG: --vmodule="" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923243 4868 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923248 4868 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923374 4868 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923379 4868 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923383 4868 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923387 4868 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923391 4868 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923395 4868 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923400 4868 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923404 4868 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923409 4868 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923413 4868 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923417 4868 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923421 4868 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923427 4868 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923432 4868 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923437 4868 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923441 4868 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923445 4868 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923449 4868 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923453 4868 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923457 4868 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923460 4868 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923464 4868 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923469 4868 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923473 4868 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923478 4868 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923483 4868 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923487 4868 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923491 4868 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923495 4868 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923498 4868 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923502 4868 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923506 4868 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923512 4868 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923516 4868 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923520 4868 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923523 4868 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923527 4868 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923530 4868 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923534 4868 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923537 4868 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923541 4868 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923544 4868 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923548 4868 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923551 4868 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923555 4868 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923558 4868 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923562 4868 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923566 4868 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923570 4868 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923573 4868 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923576 4868 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923580 4868 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923584 4868 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923587 4868 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923591 4868 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923595 4868 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923598 4868 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923602 4868 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923606 4868 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923610 4868 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923613 4868 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923620 4868 feature_gate.go:330] unrecognized feature gate: Example Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923624 4868 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923628 4868 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923635 4868 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923639 4868 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923643 4868 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923647 4868 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923651 4868 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923655 4868 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.923659 4868 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.923665 4868 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.936654 4868 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.936719 4868 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.936847 4868 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.936898 4868 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.936905 4868 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.936914 4868 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.936922 4868 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.936928 4868 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.936934 4868 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.936966 4868 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.936974 4868 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.936982 4868 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.936990 4868 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.936996 4868 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937003 4868 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937009 4868 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937017 4868 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937029 4868 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937037 4868 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937044 4868 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937052 4868 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937059 4868 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937065 4868 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937071 4868 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937077 4868 feature_gate.go:330] unrecognized feature gate: Example Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937083 4868 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937088 4868 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937094 4868 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937099 4868 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937105 4868 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937112 4868 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937120 4868 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937126 4868 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937143 4868 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937150 4868 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937156 4868 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937161 4868 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937167 4868 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937173 4868 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937180 4868 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937188 4868 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937194 4868 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937203 4868 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937210 4868 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937217 4868 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937222 4868 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937229 4868 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937235 4868 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937240 4868 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937246 4868 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937252 4868 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937257 4868 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937263 4868 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937270 4868 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937275 4868 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937281 4868 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937286 4868 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937294 4868 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937300 4868 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937305 4868 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937311 4868 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937317 4868 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937323 4868 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937328 4868 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937334 4868 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937339 4868 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937344 4868 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937350 4868 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937355 4868 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937361 4868 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937366 4868 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937372 4868 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937377 4868 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.937389 4868 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937630 4868 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937643 4868 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937649 4868 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937655 4868 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937661 4868 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937667 4868 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937672 4868 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937678 4868 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937683 4868 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937691 4868 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937700 4868 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937706 4868 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937713 4868 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937719 4868 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937724 4868 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937731 4868 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937738 4868 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937743 4868 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937749 4868 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937754 4868 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937760 4868 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937765 4868 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937771 4868 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937776 4868 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937782 4868 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937787 4868 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937794 4868 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937801 4868 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937807 4868 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937813 4868 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937819 4868 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937825 4868 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937832 4868 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937838 4868 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937844 4868 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937850 4868 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937855 4868 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937863 4868 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937868 4868 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937874 4868 feature_gate.go:330] unrecognized feature gate: Example Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937880 4868 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937888 4868 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937893 4868 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937898 4868 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937904 4868 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937909 4868 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937915 4868 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937921 4868 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937926 4868 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937932 4868 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937960 4868 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937966 4868 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937971 4868 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937977 4868 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937983 4868 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937989 4868 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.937995 4868 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.938003 4868 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.938010 4868 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.938017 4868 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.938024 4868 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.938031 4868 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.938038 4868 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.938044 4868 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.938051 4868 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.938057 4868 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.938065 4868 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.938073 4868 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.938080 4868 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.938087 4868 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 17:25:27 crc kubenswrapper[4868]: W1201 17:25:27.938095 4868 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.938106 4868 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.938396 4868 server.go:940] "Client rotation is on, will bootstrap in background" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.942590 4868 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.942974 4868 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.943692 4868 server.go:997] "Starting client certificate rotation" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.943732 4868 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.943997 4868 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-16 22:01:04.010110678 +0000 UTC Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.944154 4868 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 364h35m36.065962795s for next certificate rotation Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.963404 4868 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.965772 4868 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 17:25:27 crc kubenswrapper[4868]: I1201 17:25:27.985239 4868 log.go:25] "Validated CRI v1 runtime API" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.002328 4868 log.go:25] "Validated CRI v1 image API" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.004740 4868 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.008531 4868 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-01-17-20-33-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.008567 4868 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.021901 4868 manager.go:217] Machine: {Timestamp:2025-12-01 17:25:28.021000649 +0000 UTC m=+0.392111080 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:ba6f0959-f434-42df-bfe4-56be307a3fb1 BootID:03b0a5d8-65dc-4356-bde6-5907df7dc763 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:d4:00:70 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:d4:00:70 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:12:aa:fa Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:ed:14:0f Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:e4:66:2e Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:e1:da:bb Speed:-1 Mtu:1496} {Name:eth10 MacAddress:12:02:8a:92:95:7d Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:56:79:9c:fd:71:71 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.022145 4868 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.022291 4868 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.023039 4868 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.023228 4868 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.023265 4868 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.024231 4868 topology_manager.go:138] "Creating topology manager with none policy" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.024446 4868 container_manager_linux.go:303] "Creating device plugin manager" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.024625 4868 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.024716 4868 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.025170 4868 state_mem.go:36] "Initialized new in-memory state store" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.032920 4868 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.036064 4868 kubelet.go:418] "Attempting to sync node with API server" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.036165 4868 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.036222 4868 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.036250 4868 kubelet.go:324] "Adding apiserver pod source" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.036273 4868 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.039144 4868 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.039824 4868 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.041701 4868 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 01 17:25:28 crc kubenswrapper[4868]: W1201 17:25:28.042011 4868 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.042196 4868 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Dec 01 17:25:28 crc kubenswrapper[4868]: W1201 17:25:28.042022 4868 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.042532 4868 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.042699 4868 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.042746 4868 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.042764 4868 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.042778 4868 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.042803 4868 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.042818 4868 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.042832 4868 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.042855 4868 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.042871 4868 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.042888 4868 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.042912 4868 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.042930 4868 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.043011 4868 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.043897 4868 server.go:1280] "Started kubelet" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.044199 4868 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.044304 4868 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.045362 4868 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.046158 4868 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.046486 4868 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.046563 4868 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 01 17:25:28 crc systemd[1]: Started Kubernetes Kubelet. Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.046765 4868 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.047976 4868 server.go:460] "Adding debug handlers to kubelet server" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.048086 4868 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 08:21:55.259754088 +0000 UTC Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.049363 4868 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 806h56m27.210403176s for next certificate rotation Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.049469 4868 factory.go:55] Registering systemd factory Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.049498 4868 factory.go:221] Registration of the systemd container factory successfully Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.049677 4868 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.049647 4868 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" interval="200ms" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.049991 4868 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.050117 4868 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.050220 4868 factory.go:153] Registering CRI-O factory Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.050268 4868 factory.go:221] Registration of the crio container factory successfully Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.050443 4868 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.050528 4868 factory.go:103] Registering Raw factory Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.050575 4868 manager.go:1196] Started watching for new ooms in manager Dec 01 17:25:28 crc kubenswrapper[4868]: W1201 17:25:28.051573 4868 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.051742 4868 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.052447 4868 manager.go:319] Starting recovery of all containers Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.073401 4868 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.103:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187d2761582e3b5b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 17:25:28.043838299 +0000 UTC m=+0.414948750,LastTimestamp:2025-12-01 17:25:28.043838299 +0000 UTC m=+0.414948750,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076508 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076571 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076588 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076603 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076617 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076629 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076643 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076659 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076677 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076693 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076713 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076731 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076748 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076766 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076780 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076804 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076820 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076835 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076850 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076881 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076896 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076914 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076927 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076958 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.076994 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077016 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077058 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077072 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077087 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077102 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077119 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077134 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077153 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077168 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077183 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077223 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077239 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077255 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077269 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077281 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077297 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077313 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077329 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077350 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077375 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077392 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077408 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077423 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077452 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077468 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077483 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077497 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077518 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077534 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077550 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077594 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077612 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077625 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077637 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077649 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077666 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077680 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077697 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077713 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077738 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077775 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077788 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077799 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077811 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077822 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077836 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077848 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077861 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077875 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077892 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077908 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077920 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077931 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077964 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077975 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.077992 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078004 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078016 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078028 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078044 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078075 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078088 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078106 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078117 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078128 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078145 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078156 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078167 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078180 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078191 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078203 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078215 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078227 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078238 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078249 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078261 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078296 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078308 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078323 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078346 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078359 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078372 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078384 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078397 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078409 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078421 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078434 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078446 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078457 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078469 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078478 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078487 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078497 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078510 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078520 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078530 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078541 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078552 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078563 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078574 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078583 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078592 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078601 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078613 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078624 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078634 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078645 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078656 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078665 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078675 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078684 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078694 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078704 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078715 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078724 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078733 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078743 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078751 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078761 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078770 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078783 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078791 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078799 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078808 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078817 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078828 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078838 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078846 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078858 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078868 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078884 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078894 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078903 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078912 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078922 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078953 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078965 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078975 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078986 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.078997 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079008 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079067 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079079 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079088 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079099 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079108 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079119 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079127 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079137 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079147 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079157 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079172 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079182 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079197 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079208 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079217 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079226 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079239 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079248 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.079257 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.080976 4868 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081027 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081046 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081068 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081087 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081103 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081119 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081137 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081159 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081174 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081187 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081201 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081220 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081234 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081252 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081270 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081290 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081303 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081316 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081331 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081344 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081357 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081372 4868 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081384 4868 reconstruct.go:97] "Volume reconstruction finished" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.081393 4868 reconciler.go:26] "Reconciler: start to sync state" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.089540 4868 manager.go:324] Recovery completed Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.098577 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.104374 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.104437 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.104454 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.105524 4868 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.105549 4868 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.105572 4868 state_mem.go:36] "Initialized new in-memory state store" Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.147012 4868 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.165828 4868 policy_none.go:49] "None policy: Start" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.166867 4868 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.166909 4868 state_mem.go:35] "Initializing new in-memory state store" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.167047 4868 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.170550 4868 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.170614 4868 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.170657 4868 kubelet.go:2335] "Starting kubelet main sync loop" Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.170716 4868 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 01 17:25:28 crc kubenswrapper[4868]: W1201 17:25:28.173812 4868 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.173929 4868 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.247280 4868 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.247397 4868 manager.go:334] "Starting Device Plugin manager" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.247538 4868 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.247553 4868 server.go:79] "Starting device plugin registration server" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.248031 4868 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.248053 4868 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.248232 4868 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.248422 4868 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.248439 4868 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.250790 4868 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" interval="400ms" Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.256545 4868 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.270972 4868 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.271091 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.272069 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.272109 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.272124 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.272323 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.272493 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.272545 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.273301 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.273333 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.273343 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.273491 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.273697 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.273752 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.273809 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.273847 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.273861 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.274250 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.274288 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.274304 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.274477 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.274637 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.274650 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.274669 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.274682 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.274692 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.275510 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.275550 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.275570 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.275715 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.275745 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.275759 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.275772 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.275854 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.275900 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.276928 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.276985 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.277004 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.277033 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.277052 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.277062 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.277287 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.277319 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.277975 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.278004 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.278013 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.348603 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.349622 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.349649 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.349660 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.349683 4868 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.350836 4868 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.103:6443: connect: connection refused" node="crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.385805 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.385857 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.385883 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.385907 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.385927 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.386011 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.386059 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.386101 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.386183 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.386228 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.386262 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.386298 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.386319 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.386355 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.386377 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.391284 4868 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.103:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187d2761582e3b5b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 17:25:28.043838299 +0000 UTC m=+0.414948750,LastTimestamp:2025-12-01 17:25:28.043838299 +0000 UTC m=+0.414948750,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487357 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487426 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487453 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487508 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487535 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487559 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487576 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487655 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487608 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487750 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487710 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487584 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487928 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487988 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487998 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487709 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.488018 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.487767 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.488085 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.488116 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.488120 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.488158 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.488205 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.488237 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.488242 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.488252 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.488263 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.488306 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.488353 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.488391 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.551280 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.552684 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.552722 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.552734 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.552762 4868 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.553290 4868 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.103:6443: connect: connection refused" node="crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.612412 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.624561 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: W1201 17:25:28.639205 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-dfa0ad142726776241f15d9bd16d9398ce41ebc664c3e59e88c538ae91be2465 WatchSource:0}: Error finding container dfa0ad142726776241f15d9bd16d9398ce41ebc664c3e59e88c538ae91be2465: Status 404 returned error can't find the container with id dfa0ad142726776241f15d9bd16d9398ce41ebc664c3e59e88c538ae91be2465 Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.648620 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.651534 4868 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" interval="800ms" Dec 01 17:25:28 crc kubenswrapper[4868]: W1201 17:25:28.651622 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-3861cf931fec5cb27a820c4cfbf88e3a3f3653db46803211903f96195f92b3e8 WatchSource:0}: Error finding container 3861cf931fec5cb27a820c4cfbf88e3a3f3653db46803211903f96195f92b3e8: Status 404 returned error can't find the container with id 3861cf931fec5cb27a820c4cfbf88e3a3f3653db46803211903f96195f92b3e8 Dec 01 17:25:28 crc kubenswrapper[4868]: W1201 17:25:28.670203 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-6d21f27f6e08274746572d9b78afec91699ba421c2ae733ef61f0e22a815166e WatchSource:0}: Error finding container 6d21f27f6e08274746572d9b78afec91699ba421c2ae733ef61f0e22a815166e: Status 404 returned error can't find the container with id 6d21f27f6e08274746572d9b78afec91699ba421c2ae733ef61f0e22a815166e Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.678063 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.685995 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 17:25:28 crc kubenswrapper[4868]: W1201 17:25:28.692270 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-9740d74059bc154548bb592393e4ca4d7fd8de78b1fe2a97435dcfce15eff28c WatchSource:0}: Error finding container 9740d74059bc154548bb592393e4ca4d7fd8de78b1fe2a97435dcfce15eff28c: Status 404 returned error can't find the container with id 9740d74059bc154548bb592393e4ca4d7fd8de78b1fe2a97435dcfce15eff28c Dec 01 17:25:28 crc kubenswrapper[4868]: W1201 17:25:28.697641 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-810334c0db33835df4a01c56ffffd53543fa7d98f1c231c091a15a314f4c55a4 WatchSource:0}: Error finding container 810334c0db33835df4a01c56ffffd53543fa7d98f1c231c091a15a314f4c55a4: Status 404 returned error can't find the container with id 810334c0db33835df4a01c56ffffd53543fa7d98f1c231c091a15a314f4c55a4 Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.953559 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.955300 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.955337 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.955349 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:28 crc kubenswrapper[4868]: I1201 17:25:28.955375 4868 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.955766 4868 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.103:6443: connect: connection refused" node="crc" Dec 01 17:25:28 crc kubenswrapper[4868]: W1201 17:25:28.981099 4868 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Dec 01 17:25:28 crc kubenswrapper[4868]: E1201 17:25:28.981187 4868 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.047803 4868 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.181356 4868 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066" exitCode=0 Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.181456 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066"} Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.181622 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"810334c0db33835df4a01c56ffffd53543fa7d98f1c231c091a15a314f4c55a4"} Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.181821 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.182931 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6"} Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.182979 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9740d74059bc154548bb592393e4ca4d7fd8de78b1fe2a97435dcfce15eff28c"} Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.183036 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.183069 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.183083 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.184731 4868 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9" exitCode=0 Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.184825 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9"} Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.184874 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6d21f27f6e08274746572d9b78afec91699ba421c2ae733ef61f0e22a815166e"} Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.185028 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.186086 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.186116 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.186127 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.186914 4868 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9e2784223c59b86dc4b7b2014e7a88abb3b1ecea6596cce20690e2ac6def7fc4" exitCode=0 Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.186961 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9e2784223c59b86dc4b7b2014e7a88abb3b1ecea6596cce20690e2ac6def7fc4"} Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.186992 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3861cf931fec5cb27a820c4cfbf88e3a3f3653db46803211903f96195f92b3e8"} Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.187117 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.187725 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.187761 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.187773 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.188583 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.189266 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.189293 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.189305 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.189533 4868 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="8cca7f9573182f19dae9f87074d72f5ea36842aea72907153a7789b33eaee86b" exitCode=0 Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.189586 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"8cca7f9573182f19dae9f87074d72f5ea36842aea72907153a7789b33eaee86b"} Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.189612 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"dfa0ad142726776241f15d9bd16d9398ce41ebc664c3e59e88c538ae91be2465"} Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.189696 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.190508 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.190532 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.190541 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:29 crc kubenswrapper[4868]: W1201 17:25:29.227381 4868 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Dec 01 17:25:29 crc kubenswrapper[4868]: E1201 17:25:29.227487 4868 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Dec 01 17:25:29 crc kubenswrapper[4868]: W1201 17:25:29.244421 4868 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Dec 01 17:25:29 crc kubenswrapper[4868]: E1201 17:25:29.244540 4868 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Dec 01 17:25:29 crc kubenswrapper[4868]: E1201 17:25:29.452914 4868 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" interval="1.6s" Dec 01 17:25:29 crc kubenswrapper[4868]: W1201 17:25:29.642190 4868 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Dec 01 17:25:29 crc kubenswrapper[4868]: E1201 17:25:29.642319 4868 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.103:6443: connect: connection refused" logger="UnhandledError" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.755904 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.757582 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.757625 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.757637 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:29 crc kubenswrapper[4868]: I1201 17:25:29.757666 4868 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 17:25:29 crc kubenswrapper[4868]: E1201 17:25:29.758179 4868 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.103:6443: connect: connection refused" node="crc" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.047503 4868 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.103:6443: connect: connection refused Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.196796 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"865f0cda78b8e8ef0cd3d52fd8c4baa2e0c0f67e9d2d89752ac7a7f5e5df0a32"} Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.196858 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"120b3cade14d3c00f0a22c5f4db0c5500a669bb07a267b65bf7a386938467f16"} Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.196873 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"67c84b10381473d660a8159f66c6c2f012e0d46e588ef7f19a0f6950edc94806"} Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.197005 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.198120 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.198177 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.198193 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.202252 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.202237 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1"} Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.202405 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c"} Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.202426 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804"} Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.203392 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.203427 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.203441 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.206955 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152"} Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.207023 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e"} Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.207041 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9"} Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.207052 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7"} Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.209112 4868 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="63a347dbbb8107c0ec623b384cbad1b97a45dfa35b46862759d9472cef111f3d" exitCode=0 Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.209179 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"63a347dbbb8107c0ec623b384cbad1b97a45dfa35b46862759d9472cef111f3d"} Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.209397 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.210635 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.210665 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.210675 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.212303 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"9b1280eab687f307ac45cc184525a9aa1737ef93d7b0b1921dcb474c0d203f7d"} Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.212389 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.213217 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.213239 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.213249 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:30 crc kubenswrapper[4868]: I1201 17:25:30.674477 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.219494 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc"} Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.219676 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.221036 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.221084 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.221107 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.222548 4868 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6e752fe0c5e7f35cd10f8f0cfcc7b5dac35ee28a56a6ed321264a1c9f57d7501" exitCode=0 Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.222660 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6e752fe0c5e7f35cd10f8f0cfcc7b5dac35ee28a56a6ed321264a1c9f57d7501"} Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.222733 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.223006 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.228149 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.228210 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.228231 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.229084 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.229167 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.229194 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.358321 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.360023 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.360085 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.360103 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:31 crc kubenswrapper[4868]: I1201 17:25:31.360138 4868 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 17:25:32 crc kubenswrapper[4868]: I1201 17:25:32.237833 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6baca30001f9f6a2d8a456f42491c690459a5b6e49100fb0a247cc4fead27f3e"} Dec 01 17:25:32 crc kubenswrapper[4868]: I1201 17:25:32.237913 4868 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 17:25:32 crc kubenswrapper[4868]: I1201 17:25:32.238002 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f5ef626bc6b824e7797ec34795555859457ac79370931f06a7bda716070d65b7"} Dec 01 17:25:32 crc kubenswrapper[4868]: I1201 17:25:32.237912 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:32 crc kubenswrapper[4868]: I1201 17:25:32.238054 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"51fafa7add0d8698b2c1b1884c153f5b3e2b9ef151b1bdcbc75fac9d58599900"} Dec 01 17:25:32 crc kubenswrapper[4868]: I1201 17:25:32.238087 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b8c4df4f671c9dfd0cd8db820e4c6a75746972797cf81dd0f7fdee21a0f48564"} Dec 01 17:25:32 crc kubenswrapper[4868]: I1201 17:25:32.238031 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:32 crc kubenswrapper[4868]: I1201 17:25:32.239056 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:32 crc kubenswrapper[4868]: I1201 17:25:32.239087 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:32 crc kubenswrapper[4868]: I1201 17:25:32.239099 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:32 crc kubenswrapper[4868]: I1201 17:25:32.240164 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:32 crc kubenswrapper[4868]: I1201 17:25:32.240191 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:32 crc kubenswrapper[4868]: I1201 17:25:32.240202 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.003480 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.003784 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.005649 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.005731 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.005746 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.246326 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"dcb98f9fba578a348193332ca697b6bca447a0c0727649fbcc4d1c0b9723275c"} Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.246546 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.247702 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.247755 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.247771 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.568236 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.568481 4868 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.568543 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.570427 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.570491 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.570515 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.627170 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.674797 4868 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 01 17:25:33 crc kubenswrapper[4868]: I1201 17:25:33.674925 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 01 17:25:34 crc kubenswrapper[4868]: I1201 17:25:34.249749 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:34 crc kubenswrapper[4868]: I1201 17:25:34.249749 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:34 crc kubenswrapper[4868]: I1201 17:25:34.251697 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:34 crc kubenswrapper[4868]: I1201 17:25:34.251777 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:34 crc kubenswrapper[4868]: I1201 17:25:34.251804 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:34 crc kubenswrapper[4868]: I1201 17:25:34.251725 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:34 crc kubenswrapper[4868]: I1201 17:25:34.251856 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:34 crc kubenswrapper[4868]: I1201 17:25:34.251880 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:34 crc kubenswrapper[4868]: I1201 17:25:34.280765 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:34 crc kubenswrapper[4868]: I1201 17:25:34.521214 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:34 crc kubenswrapper[4868]: I1201 17:25:34.521492 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:34 crc kubenswrapper[4868]: I1201 17:25:34.525829 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:34 crc kubenswrapper[4868]: I1201 17:25:34.525882 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:34 crc kubenswrapper[4868]: I1201 17:25:34.525906 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:34 crc kubenswrapper[4868]: I1201 17:25:34.615716 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 01 17:25:35 crc kubenswrapper[4868]: I1201 17:25:35.254197 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:35 crc kubenswrapper[4868]: I1201 17:25:35.254259 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:35 crc kubenswrapper[4868]: I1201 17:25:35.256200 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:35 crc kubenswrapper[4868]: I1201 17:25:35.256282 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:35 crc kubenswrapper[4868]: I1201 17:25:35.256309 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:35 crc kubenswrapper[4868]: I1201 17:25:35.256375 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:35 crc kubenswrapper[4868]: I1201 17:25:35.256409 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:35 crc kubenswrapper[4868]: I1201 17:25:35.256424 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:35 crc kubenswrapper[4868]: I1201 17:25:35.752403 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 01 17:25:36 crc kubenswrapper[4868]: I1201 17:25:36.257432 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:36 crc kubenswrapper[4868]: I1201 17:25:36.258877 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:36 crc kubenswrapper[4868]: I1201 17:25:36.259024 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:36 crc kubenswrapper[4868]: I1201 17:25:36.259057 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:38 crc kubenswrapper[4868]: I1201 17:25:38.249052 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:38 crc kubenswrapper[4868]: I1201 17:25:38.249353 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:38 crc kubenswrapper[4868]: I1201 17:25:38.250808 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:38 crc kubenswrapper[4868]: I1201 17:25:38.250853 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:38 crc kubenswrapper[4868]: I1201 17:25:38.250867 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:38 crc kubenswrapper[4868]: I1201 17:25:38.255918 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:38 crc kubenswrapper[4868]: E1201 17:25:38.256646 4868 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 17:25:38 crc kubenswrapper[4868]: I1201 17:25:38.261163 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:38 crc kubenswrapper[4868]: I1201 17:25:38.261301 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:38 crc kubenswrapper[4868]: I1201 17:25:38.262233 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:38 crc kubenswrapper[4868]: I1201 17:25:38.262287 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:38 crc kubenswrapper[4868]: I1201 17:25:38.262307 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:39 crc kubenswrapper[4868]: I1201 17:25:39.264055 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:39 crc kubenswrapper[4868]: I1201 17:25:39.265377 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:39 crc kubenswrapper[4868]: I1201 17:25:39.265454 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:39 crc kubenswrapper[4868]: I1201 17:25:39.265477 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:39 crc kubenswrapper[4868]: I1201 17:25:39.269320 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:40 crc kubenswrapper[4868]: I1201 17:25:40.266363 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:40 crc kubenswrapper[4868]: I1201 17:25:40.267618 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:40 crc kubenswrapper[4868]: I1201 17:25:40.267670 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:40 crc kubenswrapper[4868]: I1201 17:25:40.267682 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:41 crc kubenswrapper[4868]: I1201 17:25:41.048206 4868 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 01 17:25:41 crc kubenswrapper[4868]: E1201 17:25:41.056363 4868 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 01 17:25:41 crc kubenswrapper[4868]: W1201 17:25:41.313393 4868 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 01 17:25:41 crc kubenswrapper[4868]: I1201 17:25:41.313502 4868 trace.go:236] Trace[1792709033]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 17:25:31.311) (total time: 10001ms): Dec 01 17:25:41 crc kubenswrapper[4868]: Trace[1792709033]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (17:25:41.313) Dec 01 17:25:41 crc kubenswrapper[4868]: Trace[1792709033]: [10.001471144s] [10.001471144s] END Dec 01 17:25:41 crc kubenswrapper[4868]: E1201 17:25:41.313527 4868 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 01 17:25:41 crc kubenswrapper[4868]: E1201 17:25:41.361719 4868 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 01 17:25:41 crc kubenswrapper[4868]: I1201 17:25:41.505297 4868 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 01 17:25:41 crc kubenswrapper[4868]: I1201 17:25:41.505382 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 01 17:25:41 crc kubenswrapper[4868]: I1201 17:25:41.516828 4868 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 01 17:25:41 crc kubenswrapper[4868]: I1201 17:25:41.516899 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 01 17:25:43 crc kubenswrapper[4868]: I1201 17:25:43.575793 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:43 crc kubenswrapper[4868]: I1201 17:25:43.576065 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:43 crc kubenswrapper[4868]: I1201 17:25:43.578424 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:43 crc kubenswrapper[4868]: I1201 17:25:43.578505 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:43 crc kubenswrapper[4868]: I1201 17:25:43.578520 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:43 crc kubenswrapper[4868]: I1201 17:25:43.581265 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:43 crc kubenswrapper[4868]: I1201 17:25:43.675313 4868 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 01 17:25:43 crc kubenswrapper[4868]: I1201 17:25:43.675458 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 17:25:44 crc kubenswrapper[4868]: I1201 17:25:44.277370 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:44 crc kubenswrapper[4868]: I1201 17:25:44.278834 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:44 crc kubenswrapper[4868]: I1201 17:25:44.278915 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:44 crc kubenswrapper[4868]: I1201 17:25:44.278939 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:44 crc kubenswrapper[4868]: I1201 17:25:44.562650 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:44 crc kubenswrapper[4868]: I1201 17:25:44.564038 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:44 crc kubenswrapper[4868]: I1201 17:25:44.564075 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:44 crc kubenswrapper[4868]: I1201 17:25:44.564085 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:44 crc kubenswrapper[4868]: I1201 17:25:44.564111 4868 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 17:25:44 crc kubenswrapper[4868]: E1201 17:25:44.568704 4868 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 01 17:25:45 crc kubenswrapper[4868]: I1201 17:25:45.771854 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 01 17:25:45 crc kubenswrapper[4868]: I1201 17:25:45.772087 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:45 crc kubenswrapper[4868]: I1201 17:25:45.773455 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:45 crc kubenswrapper[4868]: I1201 17:25:45.773498 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:45 crc kubenswrapper[4868]: I1201 17:25:45.773570 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:45 crc kubenswrapper[4868]: I1201 17:25:45.781973 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.283132 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.284418 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.284507 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.284525 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.506499 4868 trace.go:236] Trace[1534823597]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 17:25:31.777) (total time: 14724ms): Dec 01 17:25:46 crc kubenswrapper[4868]: Trace[1534823597]: ---"Objects listed" error: 14724ms (17:25:46.501) Dec 01 17:25:46 crc kubenswrapper[4868]: Trace[1534823597]: [14.72416369s] [14.72416369s] END Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.506547 4868 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.509412 4868 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.509759 4868 trace.go:236] Trace[1205729588]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 17:25:32.359) (total time: 14149ms): Dec 01 17:25:46 crc kubenswrapper[4868]: Trace[1205729588]: ---"Objects listed" error: 14149ms (17:25:46.509) Dec 01 17:25:46 crc kubenswrapper[4868]: Trace[1205729588]: [14.149768064s] [14.149768064s] END Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.509812 4868 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.510030 4868 trace.go:236] Trace[1362605774]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 17:25:31.511) (total time: 14998ms): Dec 01 17:25:46 crc kubenswrapper[4868]: Trace[1362605774]: ---"Objects listed" error: 14998ms (17:25:46.509) Dec 01 17:25:46 crc kubenswrapper[4868]: Trace[1362605774]: [14.998524645s] [14.998524645s] END Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.510054 4868 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.545795 4868 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:41980->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.545856 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:41980->192.168.126.11:17697: read: connection reset by peer" Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.546143 4868 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.546165 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.546390 4868 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 01 17:25:46 crc kubenswrapper[4868]: I1201 17:25:46.546459 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.047925 4868 apiserver.go:52] "Watching apiserver" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.051587 4868 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.051857 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.052551 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.052609 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.052779 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.053154 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.053307 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.053045 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.053527 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.053930 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.054013 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.055524 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.055833 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.056005 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.056337 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.056565 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.056622 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.056746 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.056814 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.057158 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.087500 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.103396 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.116007 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.126484 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.135621 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.146335 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.150459 4868 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.158281 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.172357 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213214 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213267 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213289 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213306 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213323 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213343 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213360 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213380 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213399 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213417 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213437 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213453 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213469 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213486 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213501 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213523 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213568 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213588 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213607 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213624 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213641 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213659 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213717 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213738 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213756 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213777 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213795 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213789 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213813 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213908 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213936 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213985 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.213907 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.214017 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.214059 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.214086 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.214114 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.214268 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.214448 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.214457 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.214520 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.214652 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.214688 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.214692 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.214963 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215088 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215179 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215258 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215347 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215398 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215432 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215516 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215555 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215560 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215617 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215651 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215682 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215709 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215712 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215739 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215773 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215800 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215826 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215859 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215884 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215911 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215943 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.215992 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216012 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216026 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216053 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216082 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216109 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216136 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216160 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216185 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216178 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216213 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216243 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216267 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216295 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216320 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216349 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216470 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216497 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216521 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216553 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216577 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216603 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216630 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216683 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216750 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216776 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216801 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216828 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216852 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216879 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216905 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216984 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217007 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217031 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217056 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217079 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217102 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217125 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217151 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217174 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217754 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217789 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217815 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217836 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217854 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217931 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217983 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218015 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218044 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218069 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218098 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218129 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218160 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218187 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218213 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218240 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218267 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218296 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218325 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218350 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218377 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218412 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218447 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218476 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218504 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218532 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218559 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218589 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218618 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218693 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218728 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218756 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218784 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218812 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218841 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218870 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218901 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218928 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218976 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219008 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219036 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219066 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219092 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219124 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219150 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219180 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219210 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219238 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219269 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219300 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219332 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219364 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219392 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219420 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219448 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219478 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219509 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219542 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219573 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219614 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219647 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219675 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219702 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219727 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219752 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219777 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219806 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219831 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219856 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219883 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219910 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219940 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219985 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220011 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220038 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220068 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220095 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220125 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220156 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220183 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220213 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220247 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220275 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220303 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220332 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220356 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220385 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220413 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220440 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220468 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220492 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220517 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220540 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220564 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220590 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220617 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220642 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220665 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220690 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220716 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220739 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220763 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220792 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220844 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220884 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220911 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220960 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221023 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221061 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221088 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221115 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221143 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221178 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221209 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221237 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221263 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221344 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221466 4868 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221485 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221501 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221517 4868 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221529 4868 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221544 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221557 4868 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221570 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221584 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221600 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221632 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216185 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216510 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216562 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216571 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216601 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216749 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.216898 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217081 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217936 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218219 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218333 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218396 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218420 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218443 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.217767 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.218825 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219014 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219051 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.219595 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220378 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220418 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220773 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220836 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.220885 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221092 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221056 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221197 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221650 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.221668 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:25:47.721634868 +0000 UTC m=+20.092745289 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.223796 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.222048 4868 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.225093 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.221855 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.225283 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.222072 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.222239 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.223109 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.223488 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.225438 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.225491 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.223775 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.223632 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.223841 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.225536 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.224243 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.224106 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.224806 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.224840 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.224861 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.224903 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.224930 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.222066 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.225925 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.226000 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.226169 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.226246 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.226408 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.226631 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.226755 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.226799 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.225573 4868 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.227034 4868 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.227031 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.227103 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:47.727068101 +0000 UTC m=+20.098178732 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.227242 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.227291 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.227393 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.227542 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.227699 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.227629 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.227815 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.228041 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.228076 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.228378 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.228825 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.228832 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.228963 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.229422 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.229603 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.229633 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.229700 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.230102 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.229749 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.229847 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.230618 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.230639 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.230732 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.231024 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.231173 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.231389 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.231804 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.232279 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.232286 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.232429 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.232575 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.232728 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.232795 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.233165 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.233198 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.233306 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.233336 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.233331 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.233926 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.234608 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.234640 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.234671 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.234995 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.234999 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.235236 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.235351 4868 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.235406 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.235439 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.235543 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.235661 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.235866 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.236460 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.236560 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.236860 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.237066 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.237065 4868 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.236652 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:47.736617222 +0000 UTC m=+20.107727633 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.237410 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.237555 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.238075 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.237082 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.238810 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.239416 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.241264 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.241346 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.242346 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.251734 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.251852 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.251935 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.252294 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.253050 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.253074 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.253088 4868 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.253158 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:47.753139538 +0000 UTC m=+20.124249959 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.253480 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.255676 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.256062 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.256259 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.256780 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.260595 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.260641 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.260664 4868 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.260745 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:47.760721348 +0000 UTC m=+20.131831759 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.261329 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.261470 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.261295 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.264729 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.264715 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.264756 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.264992 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.265281 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.265438 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.265502 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.266301 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.266327 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.266599 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.266638 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.266742 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.266862 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.266936 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.266992 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.267312 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.267320 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.267363 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.267662 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.267617 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.267815 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.267805 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.268070 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.268240 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.268513 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.268595 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.269144 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.268645 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.268666 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.269231 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.268716 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.268878 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.269527 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.269547 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.269588 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.269982 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.270113 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.270250 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.270283 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.270305 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.270468 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.271034 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.271092 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.270490 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.271189 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.271346 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.271383 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.271589 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.271650 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.271722 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.273062 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.290611 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.292143 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.295075 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.295548 4868 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc" exitCode=255 Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.295626 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc"} Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.304050 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.308929 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.309141 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.310278 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.310306 4868 scope.go:117] "RemoveContainer" containerID="e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.321742 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.328876 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.329054 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.333917 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334107 4868 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334137 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334150 4868 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334160 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334170 4868 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334181 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334191 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334104 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334204 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334268 4868 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334283 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334294 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334306 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334318 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334329 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334338 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334348 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334357 4868 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334367 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334378 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334388 4868 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334399 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334409 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334419 4868 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334429 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334438 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334447 4868 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334456 4868 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334466 4868 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334478 4868 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334487 4868 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334501 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334518 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334530 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334541 4868 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334552 4868 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334590 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334605 4868 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334618 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334630 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334642 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334654 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334666 4868 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334679 4868 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334693 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334710 4868 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334726 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334742 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334754 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334765 4868 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334775 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334784 4868 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334795 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334805 4868 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334815 4868 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334824 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334835 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334845 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334855 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334867 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334878 4868 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334889 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334899 4868 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334909 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334919 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334929 4868 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334956 4868 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334967 4868 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334978 4868 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334987 4868 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.334998 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335009 4868 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335021 4868 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335030 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335040 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335052 4868 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335062 4868 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335072 4868 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335081 4868 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335093 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335102 4868 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335112 4868 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335123 4868 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335133 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335142 4868 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335153 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335163 4868 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335173 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335184 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335194 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335204 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335213 4868 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335223 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335233 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335247 4868 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335257 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335267 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335277 4868 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335287 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335297 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335307 4868 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335316 4868 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335326 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335336 4868 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335345 4868 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335355 4868 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335367 4868 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335377 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335387 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335397 4868 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335407 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335417 4868 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335428 4868 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335438 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335448 4868 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335458 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335469 4868 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335479 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335490 4868 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335501 4868 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335510 4868 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335519 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335529 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335539 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335551 4868 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335561 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335575 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335588 4868 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335599 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335608 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335624 4868 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335637 4868 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335648 4868 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335658 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335668 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335678 4868 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335687 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335699 4868 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335712 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335724 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335736 4868 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335750 4868 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335762 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335774 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335786 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335797 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335808 4868 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335817 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335827 4868 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335836 4868 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335846 4868 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335860 4868 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335871 4868 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335885 4868 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335897 4868 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335870 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.335909 4868 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336153 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336168 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336182 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336193 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336207 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336218 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336229 4868 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336281 4868 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336291 4868 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336300 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336311 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336320 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336331 4868 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336342 4868 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336351 4868 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336362 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336373 4868 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336383 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336393 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336403 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336412 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336423 4868 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336435 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.336445 4868 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.349012 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.360009 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.366173 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.372501 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.375010 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 17:25:47 crc kubenswrapper[4868]: W1201 17:25:47.380284 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-2055894e8aed5e823c014977ef54c0c397c752b02ec842eec6f6c85854195d35 WatchSource:0}: Error finding container 2055894e8aed5e823c014977ef54c0c397c752b02ec842eec6f6c85854195d35: Status 404 returned error can't find the container with id 2055894e8aed5e823c014977ef54c0c397c752b02ec842eec6f6c85854195d35 Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.383011 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 17:25:47 crc kubenswrapper[4868]: W1201 17:25:47.402509 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-79950a9f95ff10ad140a7c481dbe35de42e3a1e54c57ff959548a4e616954f07 WatchSource:0}: Error finding container 79950a9f95ff10ad140a7c481dbe35de42e3a1e54c57ff959548a4e616954f07: Status 404 returned error can't find the container with id 79950a9f95ff10ad140a7c481dbe35de42e3a1e54c57ff959548a4e616954f07 Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.481833 4868 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.741813 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.741911 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.742048 4868 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.742049 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:25:48.741983319 +0000 UTC m=+21.113093730 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.742175 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:48.742164294 +0000 UTC m=+21.113274705 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.742216 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.742312 4868 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.742391 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:48.742375239 +0000 UTC m=+21.113485640 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.780065 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.843910 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:25:47 crc kubenswrapper[4868]: I1201 17:25:47.844042 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.844140 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.844189 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.844188 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.844200 4868 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.844221 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.844239 4868 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.844279 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:48.844258496 +0000 UTC m=+21.215368907 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:47 crc kubenswrapper[4868]: E1201 17:25:47.844294 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:48.844289166 +0000 UTC m=+21.215399577 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.171521 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:25:48 crc kubenswrapper[4868]: E1201 17:25:48.171762 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.176965 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.177764 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.178775 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.179515 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.180199 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.181600 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.182431 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.183577 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.184344 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.185690 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.186350 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.187646 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.188272 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.188882 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.190062 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.190809 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.192172 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.192666 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.193359 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.194444 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.195055 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.196117 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.196671 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.197873 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.198431 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.199253 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.199616 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.200489 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.200997 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.202202 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.202668 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.203881 4868 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.204003 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.205753 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.206897 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.207436 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.209317 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.210093 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.210998 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.211653 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.212733 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.212790 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.213197 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.214226 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.214862 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.215881 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.216558 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.217576 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.218113 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.219379 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.219916 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.220939 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.221467 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.222544 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.223234 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.223786 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.226572 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.239257 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.257556 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.272240 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.293888 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.301893 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.304742 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30"} Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.306265 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.307447 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"79950a9f95ff10ad140a7c481dbe35de42e3a1e54c57ff959548a4e616954f07"} Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.311988 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1"} Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.312040 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58"} Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.312052 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8c8127698dac4c6b4573bf5e0e33a6587f773015fb4755f91e51082d9e20206a"} Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.314529 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73"} Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.314568 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2055894e8aed5e823c014977ef54c0c397c752b02ec842eec6f6c85854195d35"} Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.326615 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.341926 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.353305 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.365711 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.380688 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.395739 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.411609 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.427777 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.442699 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.457875 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.474312 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.489609 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.502174 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.518145 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.762730 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.762830 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.762865 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:48 crc kubenswrapper[4868]: E1201 17:25:48.763059 4868 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:25:48 crc kubenswrapper[4868]: E1201 17:25:48.763132 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:50.763112755 +0000 UTC m=+23.134223186 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:25:48 crc kubenswrapper[4868]: E1201 17:25:48.763627 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:25:50.76361373 +0000 UTC m=+23.134724151 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:25:48 crc kubenswrapper[4868]: E1201 17:25:48.763690 4868 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:25:48 crc kubenswrapper[4868]: E1201 17:25:48.763726 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:50.763715793 +0000 UTC m=+23.134826224 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.864147 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:25:48 crc kubenswrapper[4868]: I1201 17:25:48.864199 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:25:48 crc kubenswrapper[4868]: E1201 17:25:48.864366 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:25:48 crc kubenswrapper[4868]: E1201 17:25:48.864389 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:25:48 crc kubenswrapper[4868]: E1201 17:25:48.864406 4868 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:48 crc kubenswrapper[4868]: E1201 17:25:48.864401 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:25:48 crc kubenswrapper[4868]: E1201 17:25:48.864441 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:25:48 crc kubenswrapper[4868]: E1201 17:25:48.864456 4868 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:48 crc kubenswrapper[4868]: E1201 17:25:48.864503 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:50.864484904 +0000 UTC m=+23.235595325 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:48 crc kubenswrapper[4868]: E1201 17:25:48.864529 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:50.864518685 +0000 UTC m=+23.235629106 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:49 crc kubenswrapper[4868]: I1201 17:25:49.176907 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:25:49 crc kubenswrapper[4868]: E1201 17:25:49.177213 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:25:49 crc kubenswrapper[4868]: I1201 17:25:49.177010 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:49 crc kubenswrapper[4868]: E1201 17:25:49.177453 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.171335 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:25:50 crc kubenswrapper[4868]: E1201 17:25:50.171555 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.323372 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25"} Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.343155 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.356752 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.378438 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.392915 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.408836 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.429904 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.447934 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.681824 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.687422 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.691097 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.699422 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.720226 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.737860 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.751023 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.770075 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.786168 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.794080 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.794156 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.794209 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:50 crc kubenswrapper[4868]: E1201 17:25:50.794285 4868 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:25:50 crc kubenswrapper[4868]: E1201 17:25:50.794402 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:25:54.79435898 +0000 UTC m=+27.165469411 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:25:50 crc kubenswrapper[4868]: E1201 17:25:50.794452 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:54.794435902 +0000 UTC m=+27.165546313 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:25:50 crc kubenswrapper[4868]: E1201 17:25:50.794479 4868 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:25:50 crc kubenswrapper[4868]: E1201 17:25:50.794654 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:54.794614958 +0000 UTC m=+27.165725519 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.813330 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.844106 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.866637 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.880910 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.895366 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.895439 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:25:50 crc kubenswrapper[4868]: E1201 17:25:50.895612 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:25:50 crc kubenswrapper[4868]: E1201 17:25:50.895622 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:25:50 crc kubenswrapper[4868]: E1201 17:25:50.895638 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:25:50 crc kubenswrapper[4868]: E1201 17:25:50.895651 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:25:50 crc kubenswrapper[4868]: E1201 17:25:50.895659 4868 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:50 crc kubenswrapper[4868]: E1201 17:25:50.895667 4868 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:50 crc kubenswrapper[4868]: E1201 17:25:50.895739 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:54.895715968 +0000 UTC m=+27.266826389 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:50 crc kubenswrapper[4868]: E1201 17:25:50.895766 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:54.895755959 +0000 UTC m=+27.266866390 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.903454 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.917376 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.947095 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.967304 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.969323 4868 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.970874 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.970919 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.970934 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.971043 4868 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.977964 4868 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.978294 4868 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.979595 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.979640 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.979651 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.979670 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.979681 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:50Z","lastTransitionTime":"2025-12-01T17:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:50 crc kubenswrapper[4868]: I1201 17:25:50.983553 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:50Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:51 crc kubenswrapper[4868]: E1201 17:25:51.007419 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:51Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.018220 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.018277 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.018291 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.018313 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.018328 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:51Z","lastTransitionTime":"2025-12-01T17:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:51 crc kubenswrapper[4868]: E1201 17:25:51.031308 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:51Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.042604 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.042633 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.042642 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.042656 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.042666 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:51Z","lastTransitionTime":"2025-12-01T17:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:51 crc kubenswrapper[4868]: E1201 17:25:51.062272 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:51Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.073141 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.073177 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.073187 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.073202 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.073212 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:51Z","lastTransitionTime":"2025-12-01T17:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:51 crc kubenswrapper[4868]: E1201 17:25:51.095388 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:51Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.099208 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.099245 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.099258 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.099275 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.099287 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:51Z","lastTransitionTime":"2025-12-01T17:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.171055 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.171085 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:25:51 crc kubenswrapper[4868]: E1201 17:25:51.171227 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:25:51 crc kubenswrapper[4868]: E1201 17:25:51.171381 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:25:51 crc kubenswrapper[4868]: E1201 17:25:51.370604 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:51Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:51 crc kubenswrapper[4868]: E1201 17:25:51.371442 4868 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.373001 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.373045 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.373059 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.373079 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.373092 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:51Z","lastTransitionTime":"2025-12-01T17:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.475621 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.475669 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.475678 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.475694 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.475704 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:51Z","lastTransitionTime":"2025-12-01T17:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.578174 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.578224 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.578236 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.578254 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.578266 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:51Z","lastTransitionTime":"2025-12-01T17:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.680965 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.681010 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.681022 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.681043 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.681054 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:51Z","lastTransitionTime":"2025-12-01T17:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.783421 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.783761 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.783775 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.783797 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.783811 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:51Z","lastTransitionTime":"2025-12-01T17:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.885789 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.885833 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.885856 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.885876 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.885888 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:51Z","lastTransitionTime":"2025-12-01T17:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.988049 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.988099 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.988108 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.988130 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:51 crc kubenswrapper[4868]: I1201 17:25:51.988142 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:51Z","lastTransitionTime":"2025-12-01T17:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.090566 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.090611 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.090623 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.090641 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.090654 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:52Z","lastTransitionTime":"2025-12-01T17:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.172013 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:25:52 crc kubenswrapper[4868]: E1201 17:25:52.172190 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.192556 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.192688 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.192748 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.192849 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.192917 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:52Z","lastTransitionTime":"2025-12-01T17:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.296060 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.296417 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.296561 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.296753 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.296974 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:52Z","lastTransitionTime":"2025-12-01T17:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.399454 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.400022 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.400058 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.400078 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.400090 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:52Z","lastTransitionTime":"2025-12-01T17:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.653124 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.653163 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.653173 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.653192 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.653202 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:52Z","lastTransitionTime":"2025-12-01T17:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.803471 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.803509 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.803517 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.803533 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.803543 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:52Z","lastTransitionTime":"2025-12-01T17:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.905576 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.905604 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.905612 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.905626 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:52 crc kubenswrapper[4868]: I1201 17:25:52.905635 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:52Z","lastTransitionTime":"2025-12-01T17:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.007542 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.007606 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.007636 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.007683 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.007710 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:53Z","lastTransitionTime":"2025-12-01T17:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.110621 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.110659 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.110669 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.110686 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.110700 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:53Z","lastTransitionTime":"2025-12-01T17:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.171147 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:25:53 crc kubenswrapper[4868]: E1201 17:25:53.171281 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.171718 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:53 crc kubenswrapper[4868]: E1201 17:25:53.171775 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.213340 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.213372 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.213384 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.213398 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.213408 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:53Z","lastTransitionTime":"2025-12-01T17:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.364677 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.364708 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.364717 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.364731 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.364741 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:53Z","lastTransitionTime":"2025-12-01T17:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.466645 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.466683 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.466692 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.466709 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.466719 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:53Z","lastTransitionTime":"2025-12-01T17:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.593749 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.593797 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.593810 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.593835 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.593849 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:53Z","lastTransitionTime":"2025-12-01T17:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.698398 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.698440 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.698451 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.698469 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.698492 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:53Z","lastTransitionTime":"2025-12-01T17:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.800751 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.800795 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.800807 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.800829 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.800840 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:53Z","lastTransitionTime":"2025-12-01T17:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.903738 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.903797 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.903808 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.903831 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:53 crc kubenswrapper[4868]: I1201 17:25:53.903848 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:53Z","lastTransitionTime":"2025-12-01T17:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.006213 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.006246 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.006274 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.006294 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.006308 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:54Z","lastTransitionTime":"2025-12-01T17:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.109386 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.109435 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.109445 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.109463 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.109473 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:54Z","lastTransitionTime":"2025-12-01T17:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.171654 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.171785 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.340887 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.340930 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.340952 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.340970 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.340980 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:54Z","lastTransitionTime":"2025-12-01T17:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.414367 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-h9tlw"] Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.414719 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.416620 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-c7mwj"] Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.417073 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-c7mwj" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.421559 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 17:25:54 crc kubenswrapper[4868]: W1201 17:25:54.421749 4868 reflector.go:561] object-"openshift-multus"/"cni-copy-resources": failed to list *v1.ConfigMap: configmaps "cni-copy-resources" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.421786 4868 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"cni-copy-resources\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cni-copy-resources\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.421838 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 17:25:54 crc kubenswrapper[4868]: W1201 17:25:54.421964 4868 reflector.go:561] object-"openshift-multus"/"multus-daemon-config": failed to list *v1.ConfigMap: configmaps "multus-daemon-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.421983 4868 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-daemon-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"multus-daemon-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.422150 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.422872 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.423410 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.423524 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434427 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-729qk\" (UniqueName: \"kubernetes.io/projected/76081c1f-3c70-497b-8591-53350dc17405-kube-api-access-729qk\") pod \"node-resolver-c7mwj\" (UID: \"76081c1f-3c70-497b-8591-53350dc17405\") " pod="openshift-dns/node-resolver-c7mwj" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434465 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8ce688b4-36b1-48d4-a6d1-8aec723125c4-cni-binary-copy\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434488 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-run-netns\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434507 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-var-lib-cni-bin\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434527 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-multus-socket-dir-parent\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434547 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9mv8\" (UniqueName: \"kubernetes.io/projected/8ce688b4-36b1-48d4-a6d1-8aec723125c4-kube-api-access-q9mv8\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434569 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-os-release\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434583 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-multus-conf-dir\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434598 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8ce688b4-36b1-48d4-a6d1-8aec723125c4-multus-daemon-config\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434612 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-system-cni-dir\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434626 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-multus-cni-dir\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434640 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-cnibin\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434653 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-run-k8s-cni-cncf-io\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434667 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-etc-kubernetes\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434685 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-hostroot\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434717 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-var-lib-cni-multus\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434733 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-var-lib-kubelet\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434751 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-run-multus-certs\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.434767 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/76081c1f-3c70-497b-8591-53350dc17405-hosts-file\") pod \"node-resolver-c7mwj\" (UID: \"76081c1f-3c70-497b-8591-53350dc17405\") " pod="openshift-dns/node-resolver-c7mwj" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.443179 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.443212 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.443221 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.443239 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.443250 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:54Z","lastTransitionTime":"2025-12-01T17:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535367 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-cnibin\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535424 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-run-k8s-cni-cncf-io\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535448 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-multus-conf-dir\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535476 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8ce688b4-36b1-48d4-a6d1-8aec723125c4-multus-daemon-config\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535501 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-system-cni-dir\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535521 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-cnibin\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535516 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-multus-conf-dir\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535575 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-system-cni-dir\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535516 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-run-k8s-cni-cncf-io\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535526 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-multus-cni-dir\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535639 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-etc-kubernetes\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535667 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-hostroot\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535710 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-multus-cni-dir\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535713 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-run-multus-certs\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535770 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-etc-kubernetes\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535767 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-hostroot\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535792 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-run-multus-certs\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535802 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/76081c1f-3c70-497b-8591-53350dc17405-hosts-file\") pod \"node-resolver-c7mwj\" (UID: \"76081c1f-3c70-497b-8591-53350dc17405\") " pod="openshift-dns/node-resolver-c7mwj" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535841 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-var-lib-cni-multus\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535865 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-var-lib-kubelet\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535877 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/76081c1f-3c70-497b-8591-53350dc17405-hosts-file\") pod \"node-resolver-c7mwj\" (UID: \"76081c1f-3c70-497b-8591-53350dc17405\") " pod="openshift-dns/node-resolver-c7mwj" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535886 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-var-lib-cni-multus\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535902 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-729qk\" (UniqueName: \"kubernetes.io/projected/76081c1f-3c70-497b-8591-53350dc17405-kube-api-access-729qk\") pod \"node-resolver-c7mwj\" (UID: \"76081c1f-3c70-497b-8591-53350dc17405\") " pod="openshift-dns/node-resolver-c7mwj" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535918 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-var-lib-kubelet\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535924 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-run-netns\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535958 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-var-lib-cni-bin\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535973 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-run-netns\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535983 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8ce688b4-36b1-48d4-a6d1-8aec723125c4-cni-binary-copy\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.535999 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-host-var-lib-cni-bin\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.536007 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-os-release\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.536024 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-multus-socket-dir-parent\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.536045 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9mv8\" (UniqueName: \"kubernetes.io/projected/8ce688b4-36b1-48d4-a6d1-8aec723125c4-kube-api-access-q9mv8\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.536066 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-os-release\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.536099 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8ce688b4-36b1-48d4-a6d1-8aec723125c4-multus-socket-dir-parent\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.581713 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.581751 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.581761 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.581778 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.581787 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:54Z","lastTransitionTime":"2025-12-01T17:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.620822 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:54Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.641625 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-729qk\" (UniqueName: \"kubernetes.io/projected/76081c1f-3c70-497b-8591-53350dc17405-kube-api-access-729qk\") pod \"node-resolver-c7mwj\" (UID: \"76081c1f-3c70-497b-8591-53350dc17405\") " pod="openshift-dns/node-resolver-c7mwj" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.652011 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9mv8\" (UniqueName: \"kubernetes.io/projected/8ce688b4-36b1-48d4-a6d1-8aec723125c4-kube-api-access-q9mv8\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.684267 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.684325 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.684336 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.684352 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.684363 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:54Z","lastTransitionTime":"2025-12-01T17:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.703473 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:54Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.737718 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-c7mwj" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.749824 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:54Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.793307 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.793341 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.793353 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.793371 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.793383 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:54Z","lastTransitionTime":"2025-12-01T17:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.841614 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.841746 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.841810 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.842010 4868 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.842084 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:26:02.842058466 +0000 UTC m=+35.213168877 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.842395 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:26:02.842351414 +0000 UTC m=+35.213461825 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.842522 4868 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.842572 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:26:02.842562821 +0000 UTC m=+35.213673232 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.844200 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:54Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.919234 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.919271 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.919319 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.919343 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.919356 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:54Z","lastTransitionTime":"2025-12-01T17:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.943972 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.944050 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.944213 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.944231 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.944242 4868 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.944290 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 17:26:02.944273708 +0000 UTC m=+35.315384119 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.944602 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.944623 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.944637 4868 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:54 crc kubenswrapper[4868]: E1201 17:25:54.944672 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 17:26:02.944662509 +0000 UTC m=+35.315772920 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.955037 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:54Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:54 crc kubenswrapper[4868]: I1201 17:25:54.969739 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:54Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.001164 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:54Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.021613 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.021661 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.021672 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.021690 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.021702 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:55Z","lastTransitionTime":"2025-12-01T17:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.026400 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.032886 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-hw725"] Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.033613 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.037320 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.037623 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.039157 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-q9z6l"] Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.039898 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.041474 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.041919 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-njgpn"] Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.042165 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.042402 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.043113 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.043232 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.043340 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.043441 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.044128 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.044268 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.054004 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.054042 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.054317 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.054490 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.054574 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.073237 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.087360 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.101587 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.114343 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.124142 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.124186 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.124196 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.124212 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.124222 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:55Z","lastTransitionTime":"2025-12-01T17:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.134348 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146105 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/46acee96-a86d-4abf-a79f-7cfc61fbab92-cnibin\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146170 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-log-socket\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146194 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-cni-bin\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146217 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-env-overrides\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146238 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78dkp\" (UniqueName: \"kubernetes.io/projected/ec4e08bf-df56-4490-8e79-a4175c727405-kube-api-access-78dkp\") pod \"machine-config-daemon-njgpn\" (UID: \"ec4e08bf-df56-4490-8e79-a4175c727405\") " pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146274 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-node-log\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146317 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/46acee96-a86d-4abf-a79f-7cfc61fbab92-system-cni-dir\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146438 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-systemd\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146467 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/46acee96-a86d-4abf-a79f-7cfc61fbab92-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146493 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-cni-netd\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146545 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/46acee96-a86d-4abf-a79f-7cfc61fbab92-os-release\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146587 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/46acee96-a86d-4abf-a79f-7cfc61fbab92-cni-binary-copy\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146630 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/46acee96-a86d-4abf-a79f-7cfc61fbab92-tuning-conf-dir\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146673 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2cpz\" (UniqueName: \"kubernetes.io/projected/5baca705-90cc-4198-b641-214a2b915086-kube-api-access-j2cpz\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146756 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-kubelet\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.146793 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-run-netns\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.147016 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-var-lib-openvswitch\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.147043 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-etc-openvswitch\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.147067 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-ovn\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.147133 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.147169 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ec4e08bf-df56-4490-8e79-a4175c727405-mcd-auth-proxy-config\") pod \"machine-config-daemon-njgpn\" (UID: \"ec4e08bf-df56-4490-8e79-a4175c727405\") " pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.147202 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-openvswitch\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.147238 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-systemd-units\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.147259 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-slash\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.147306 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5baca705-90cc-4198-b641-214a2b915086-ovn-node-metrics-cert\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.147330 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-ovnkube-script-lib\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.147409 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ec4e08bf-df56-4490-8e79-a4175c727405-rootfs\") pod \"machine-config-daemon-njgpn\" (UID: \"ec4e08bf-df56-4490-8e79-a4175c727405\") " pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.147437 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ec4e08bf-df56-4490-8e79-a4175c727405-proxy-tls\") pod \"machine-config-daemon-njgpn\" (UID: \"ec4e08bf-df56-4490-8e79-a4175c727405\") " pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.147459 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-run-ovn-kubernetes\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.147480 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-ovnkube-config\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.147504 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk4sq\" (UniqueName: \"kubernetes.io/projected/46acee96-a86d-4abf-a79f-7cfc61fbab92-kube-api-access-qk4sq\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.150058 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.165104 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.171224 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.171227 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:55 crc kubenswrapper[4868]: E1201 17:25:55.171346 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:25:55 crc kubenswrapper[4868]: E1201 17:25:55.171529 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.179985 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.198852 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.218382 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.226976 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.227020 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.227034 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.227054 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.227069 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:55Z","lastTransitionTime":"2025-12-01T17:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.235409 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.237216 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8ce688b4-36b1-48d4-a6d1-8aec723125c4-multus-daemon-config\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.239438 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.266818 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-var-lib-openvswitch\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.266867 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-etc-openvswitch\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.266888 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-ovn\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.266910 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.266953 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ec4e08bf-df56-4490-8e79-a4175c727405-mcd-auth-proxy-config\") pod \"machine-config-daemon-njgpn\" (UID: \"ec4e08bf-df56-4490-8e79-a4175c727405\") " pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.266986 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-ovn\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.266996 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-etc-openvswitch\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267047 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-systemd-units\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.266990 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-var-lib-openvswitch\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.266987 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-systemd-units\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267019 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267115 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-openvswitch\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267152 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-slash\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267175 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-openvswitch\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267215 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-slash\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267179 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5baca705-90cc-4198-b641-214a2b915086-ovn-node-metrics-cert\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267250 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-ovnkube-script-lib\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267275 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ec4e08bf-df56-4490-8e79-a4175c727405-rootfs\") pod \"machine-config-daemon-njgpn\" (UID: \"ec4e08bf-df56-4490-8e79-a4175c727405\") " pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267302 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ec4e08bf-df56-4490-8e79-a4175c727405-proxy-tls\") pod \"machine-config-daemon-njgpn\" (UID: \"ec4e08bf-df56-4490-8e79-a4175c727405\") " pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267328 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-run-ovn-kubernetes\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267349 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-ovnkube-config\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267357 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ec4e08bf-df56-4490-8e79-a4175c727405-rootfs\") pod \"machine-config-daemon-njgpn\" (UID: \"ec4e08bf-df56-4490-8e79-a4175c727405\") " pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267397 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk4sq\" (UniqueName: \"kubernetes.io/projected/46acee96-a86d-4abf-a79f-7cfc61fbab92-kube-api-access-qk4sq\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267434 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-log-socket\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267454 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/46acee96-a86d-4abf-a79f-7cfc61fbab92-cnibin\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267476 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-node-log\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267498 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-cni-bin\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267520 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-env-overrides\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267546 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78dkp\" (UniqueName: \"kubernetes.io/projected/ec4e08bf-df56-4490-8e79-a4175c727405-kube-api-access-78dkp\") pod \"machine-config-daemon-njgpn\" (UID: \"ec4e08bf-df56-4490-8e79-a4175c727405\") " pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267584 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/46acee96-a86d-4abf-a79f-7cfc61fbab92-system-cni-dir\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267609 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/46acee96-a86d-4abf-a79f-7cfc61fbab92-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267631 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-systemd\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267655 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-cni-netd\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267678 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/46acee96-a86d-4abf-a79f-7cfc61fbab92-os-release\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267727 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/46acee96-a86d-4abf-a79f-7cfc61fbab92-cni-binary-copy\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267763 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/46acee96-a86d-4abf-a79f-7cfc61fbab92-tuning-conf-dir\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267789 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2cpz\" (UniqueName: \"kubernetes.io/projected/5baca705-90cc-4198-b641-214a2b915086-kube-api-access-j2cpz\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267793 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-cni-bin\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267810 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-kubelet\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267828 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ec4e08bf-df56-4490-8e79-a4175c727405-mcd-auth-proxy-config\") pod \"machine-config-daemon-njgpn\" (UID: \"ec4e08bf-df56-4490-8e79-a4175c727405\") " pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267833 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-run-netns\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267863 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-run-netns\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267920 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-cni-netd\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267931 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/46acee96-a86d-4abf-a79f-7cfc61fbab92-os-release\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.267958 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-run-ovn-kubernetes\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.268033 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-kubelet\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.268124 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/46acee96-a86d-4abf-a79f-7cfc61fbab92-cnibin\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.268179 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-log-socket\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.268194 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-node-log\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.268223 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/46acee96-a86d-4abf-a79f-7cfc61fbab92-system-cni-dir\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.268271 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-systemd\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.268338 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-ovnkube-script-lib\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.268419 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/46acee96-a86d-4abf-a79f-7cfc61fbab92-tuning-conf-dir\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.268557 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-ovnkube-config\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.268791 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/46acee96-a86d-4abf-a79f-7cfc61fbab92-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.268806 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-env-overrides\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.271682 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ec4e08bf-df56-4490-8e79-a4175c727405-proxy-tls\") pod \"machine-config-daemon-njgpn\" (UID: \"ec4e08bf-df56-4490-8e79-a4175c727405\") " pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.272291 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.284556 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2cpz\" (UniqueName: \"kubernetes.io/projected/5baca705-90cc-4198-b641-214a2b915086-kube-api-access-j2cpz\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.284853 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.284880 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk4sq\" (UniqueName: \"kubernetes.io/projected/46acee96-a86d-4abf-a79f-7cfc61fbab92-kube-api-access-qk4sq\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.289559 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5baca705-90cc-4198-b641-214a2b915086-ovn-node-metrics-cert\") pod \"ovnkube-node-q9z6l\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.307296 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78dkp\" (UniqueName: \"kubernetes.io/projected/ec4e08bf-df56-4490-8e79-a4175c727405-kube-api-access-78dkp\") pod \"machine-config-daemon-njgpn\" (UID: \"ec4e08bf-df56-4490-8e79-a4175c727405\") " pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.330674 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.330743 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.330762 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.330793 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.330815 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:55Z","lastTransitionTime":"2025-12-01T17:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.346926 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-c7mwj" event={"ID":"76081c1f-3c70-497b-8591-53350dc17405","Type":"ContainerStarted","Data":"867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681"} Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.347014 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-c7mwj" event={"ID":"76081c1f-3c70-497b-8591-53350dc17405","Type":"ContainerStarted","Data":"8b72defcb1ea2ff0ad58ac4e05aa8380d7f0057735d587c64bd7a6e5613fe72c"} Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.362889 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.366102 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.374084 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.379832 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.391959 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.407410 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.425485 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.433849 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.433904 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.433923 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.433966 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.433980 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:55Z","lastTransitionTime":"2025-12-01T17:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.440100 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.459028 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.474202 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.489535 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.506336 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.519336 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.532229 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.535814 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.535871 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.535886 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.535904 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.535917 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:55Z","lastTransitionTime":"2025-12-01T17:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:55 crc kubenswrapper[4868]: E1201 17:25:55.536154 4868 configmap.go:193] Couldn't get configMap openshift-multus/cni-copy-resources: failed to sync configmap cache: timed out waiting for the condition Dec 01 17:25:55 crc kubenswrapper[4868]: E1201 17:25:55.536220 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8ce688b4-36b1-48d4-a6d1-8aec723125c4-cni-binary-copy podName:8ce688b4-36b1-48d4-a6d1-8aec723125c4 nodeName:}" failed. No retries permitted until 2025-12-01 17:25:56.036200927 +0000 UTC m=+28.407311338 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-binary-copy" (UniqueName: "kubernetes.io/configmap/8ce688b4-36b1-48d4-a6d1-8aec723125c4-cni-binary-copy") pod "multus-h9tlw" (UID: "8ce688b4-36b1-48d4-a6d1-8aec723125c4") : failed to sync configmap cache: timed out waiting for the condition Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.543553 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.638841 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.638881 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.638891 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.638909 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.638922 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:55Z","lastTransitionTime":"2025-12-01T17:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.742326 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.742366 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.742376 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.742393 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.742404 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:55Z","lastTransitionTime":"2025-12-01T17:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.814022 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.819096 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/46acee96-a86d-4abf-a79f-7cfc61fbab92-cni-binary-copy\") pod \"multus-additional-cni-plugins-hw725\" (UID: \"46acee96-a86d-4abf-a79f-7cfc61fbab92\") " pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.844962 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.845007 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.845018 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.845038 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.845052 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:55Z","lastTransitionTime":"2025-12-01T17:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.948711 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.948762 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.948778 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.948803 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.948821 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:55Z","lastTransitionTime":"2025-12-01T17:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:55 crc kubenswrapper[4868]: I1201 17:25:55.956330 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-hw725" Dec 01 17:25:55 crc kubenswrapper[4868]: W1201 17:25:55.974176 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46acee96_a86d_4abf_a79f_7cfc61fbab92.slice/crio-1c7b62db9181b3d71cdc19ede756b1014dd8b154106635924ab89281905d8244 WatchSource:0}: Error finding container 1c7b62db9181b3d71cdc19ede756b1014dd8b154106635924ab89281905d8244: Status 404 returned error can't find the container with id 1c7b62db9181b3d71cdc19ede756b1014dd8b154106635924ab89281905d8244 Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.066469 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.066513 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.066527 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.066547 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.066559 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:56Z","lastTransitionTime":"2025-12-01T17:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.091407 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8ce688b4-36b1-48d4-a6d1-8aec723125c4-cni-binary-copy\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.092290 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8ce688b4-36b1-48d4-a6d1-8aec723125c4-cni-binary-copy\") pod \"multus-h9tlw\" (UID: \"8ce688b4-36b1-48d4-a6d1-8aec723125c4\") " pod="openshift-multus/multus-h9tlw" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.168643 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.168724 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.168757 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.168793 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.168848 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:56Z","lastTransitionTime":"2025-12-01T17:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.171639 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:25:56 crc kubenswrapper[4868]: E1201 17:25:56.171743 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.229833 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-h9tlw" Dec 01 17:25:56 crc kubenswrapper[4868]: W1201 17:25:56.253044 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ce688b4_36b1_48d4_a6d1_8aec723125c4.slice/crio-89aa387ad610dd10dc93810edb9a0f9eab263c3d9666c740508161f635f1ee83 WatchSource:0}: Error finding container 89aa387ad610dd10dc93810edb9a0f9eab263c3d9666c740508161f635f1ee83: Status 404 returned error can't find the container with id 89aa387ad610dd10dc93810edb9a0f9eab263c3d9666c740508161f635f1ee83 Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.272280 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.272617 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.272631 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.272912 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.272960 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:56Z","lastTransitionTime":"2025-12-01T17:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.352202 4868 generic.go:334] "Generic (PLEG): container finished" podID="5baca705-90cc-4198-b641-214a2b915086" containerID="e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5" exitCode=0 Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.352313 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerDied","Data":"e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.352345 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerStarted","Data":"ef85c890c89cf0cd7e1efcda5f1342842e1535dfceed118f62c4c9af9c11bb55"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.354574 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.354618 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.354629 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"c940cfe0f2c72f6dd76e583186a1c69c0010fda16995c0847e589c755f710205"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.364204 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9tlw" event={"ID":"8ce688b4-36b1-48d4-a6d1-8aec723125c4","Type":"ContainerStarted","Data":"b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.364289 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9tlw" event={"ID":"8ce688b4-36b1-48d4-a6d1-8aec723125c4","Type":"ContainerStarted","Data":"89aa387ad610dd10dc93810edb9a0f9eab263c3d9666c740508161f635f1ee83"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.368187 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" event={"ID":"46acee96-a86d-4abf-a79f-7cfc61fbab92","Type":"ContainerStarted","Data":"d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.368250 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" event={"ID":"46acee96-a86d-4abf-a79f-7cfc61fbab92","Type":"ContainerStarted","Data":"1c7b62db9181b3d71cdc19ede756b1014dd8b154106635924ab89281905d8244"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.374856 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.375503 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.375528 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.375538 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.375553 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.375564 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:56Z","lastTransitionTime":"2025-12-01T17:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.389385 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.403275 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.414146 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.429106 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.471903 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.493189 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.493235 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.493249 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.493267 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.493280 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:56Z","lastTransitionTime":"2025-12-01T17:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.506238 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.518965 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.530332 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.543815 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.570994 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.582407 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.599135 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.599168 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.599176 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.599192 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.599202 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:56Z","lastTransitionTime":"2025-12-01T17:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.604925 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.663301 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.678624 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.691125 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.701405 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.701440 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.701448 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.701463 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.701475 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:56Z","lastTransitionTime":"2025-12-01T17:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.714361 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.730661 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.766198 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.780698 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.801230 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.803699 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.803740 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.803750 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.803772 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.803783 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:56Z","lastTransitionTime":"2025-12-01T17:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.816583 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.829857 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.842681 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.860126 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.873479 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:56Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.906427 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.906457 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.906465 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.906480 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:56 crc kubenswrapper[4868]: I1201 17:25:56.906489 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:56Z","lastTransitionTime":"2025-12-01T17:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.009665 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.010223 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.010232 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.010250 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.010262 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:57Z","lastTransitionTime":"2025-12-01T17:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.113842 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.113902 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.113915 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.113937 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.113983 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:57Z","lastTransitionTime":"2025-12-01T17:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.171135 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.171135 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:57 crc kubenswrapper[4868]: E1201 17:25:57.171415 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:25:57 crc kubenswrapper[4868]: E1201 17:25:57.171567 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.216894 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.216985 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.217009 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.217041 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.217061 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:57Z","lastTransitionTime":"2025-12-01T17:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.330626 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.330663 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.330671 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.330878 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.330888 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:57Z","lastTransitionTime":"2025-12-01T17:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.386015 4868 generic.go:334] "Generic (PLEG): container finished" podID="46acee96-a86d-4abf-a79f-7cfc61fbab92" containerID="d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd" exitCode=0 Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.386107 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" event={"ID":"46acee96-a86d-4abf-a79f-7cfc61fbab92","Type":"ContainerDied","Data":"d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd"} Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.393722 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerStarted","Data":"bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b"} Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.415782 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:57Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.433133 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:57Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.438255 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.438309 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.438321 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.438345 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.438359 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:57Z","lastTransitionTime":"2025-12-01T17:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.448153 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:57Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.492969 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:57Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.506769 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:57Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.517856 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:57Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.527285 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:57Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.541255 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.541295 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.541306 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.541329 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.541342 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:57Z","lastTransitionTime":"2025-12-01T17:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.554278 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:57Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.607640 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:57Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.619285 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:57Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.647606 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:57Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.653996 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.654031 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.654044 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.654061 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.654073 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:57Z","lastTransitionTime":"2025-12-01T17:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.674354 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:57Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.689145 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:57Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.756526 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.756572 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.756586 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.756609 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.756624 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:57Z","lastTransitionTime":"2025-12-01T17:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.858857 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.858906 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.858918 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.858936 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.858962 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:57Z","lastTransitionTime":"2025-12-01T17:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.986387 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.986428 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.986440 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.986461 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:57 crc kubenswrapper[4868]: I1201 17:25:57.986473 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:57Z","lastTransitionTime":"2025-12-01T17:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.090912 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.090981 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.090994 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.091016 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.091027 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:58Z","lastTransitionTime":"2025-12-01T17:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.172385 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:25:58 crc kubenswrapper[4868]: E1201 17:25:58.172525 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.197695 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.197731 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.197743 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.197759 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.197767 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:58Z","lastTransitionTime":"2025-12-01T17:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.206728 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.221898 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.314502 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.315859 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.316138 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.316150 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.316168 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.316180 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:58Z","lastTransitionTime":"2025-12-01T17:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.337315 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.359933 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.382732 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.409023 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerStarted","Data":"7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854"} Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.409073 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerStarted","Data":"df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26"} Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.409083 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerStarted","Data":"f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113"} Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.409091 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerStarted","Data":"656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df"} Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.409101 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerStarted","Data":"5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4"} Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.410431 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" event={"ID":"46acee96-a86d-4abf-a79f-7cfc61fbab92","Type":"ContainerStarted","Data":"2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246"} Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.413085 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.424910 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.424963 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.424973 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.424989 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.425000 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:58Z","lastTransitionTime":"2025-12-01T17:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.457653 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.470564 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.482707 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.492983 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.507231 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.519719 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.530582 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.530625 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.530637 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.530657 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.530669 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:58Z","lastTransitionTime":"2025-12-01T17:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.534249 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.549322 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.564847 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.580537 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.594515 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.607658 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.619938 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.632424 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.636353 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.636386 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.636397 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.636412 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.636424 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:58Z","lastTransitionTime":"2025-12-01T17:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.646803 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.677850 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.690911 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.707933 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.729277 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.743350 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.743428 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.743449 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.743485 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.743505 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:58Z","lastTransitionTime":"2025-12-01T17:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.847929 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.847993 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.848005 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.848021 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.848031 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:58Z","lastTransitionTime":"2025-12-01T17:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.963315 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.963367 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.963382 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.963404 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:58 crc kubenswrapper[4868]: I1201 17:25:58.963416 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:58Z","lastTransitionTime":"2025-12-01T17:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.066084 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.066131 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.066145 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.066165 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.066179 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:59Z","lastTransitionTime":"2025-12-01T17:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.169782 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.169846 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.169860 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.169882 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.169909 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:59Z","lastTransitionTime":"2025-12-01T17:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.172927 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.172981 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:25:59 crc kubenswrapper[4868]: E1201 17:25:59.173071 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:25:59 crc kubenswrapper[4868]: E1201 17:25:59.173180 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.272173 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.272227 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.272239 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.272258 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.272268 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:59Z","lastTransitionTime":"2025-12-01T17:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.374761 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.374795 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.374804 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.374818 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.374829 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:59Z","lastTransitionTime":"2025-12-01T17:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.416587 4868 generic.go:334] "Generic (PLEG): container finished" podID="46acee96-a86d-4abf-a79f-7cfc61fbab92" containerID="2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246" exitCode=0 Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.416663 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" event={"ID":"46acee96-a86d-4abf-a79f-7cfc61fbab92","Type":"ContainerDied","Data":"2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246"} Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.456445 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.472524 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.480673 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.480717 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.480728 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.480746 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.480761 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:59Z","lastTransitionTime":"2025-12-01T17:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.490993 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.505013 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.528104 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.546585 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.561275 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.575984 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.583307 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.583351 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.583362 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.583379 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.583392 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:59Z","lastTransitionTime":"2025-12-01T17:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.592261 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.610641 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.627124 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.648394 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.653228 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-64mjk"] Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.653699 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-64mjk" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.661575 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.661632 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.661832 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.662038 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.673117 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b184c651-661a-4d84-bb0a-75107bf5de1b-serviceca\") pod \"node-ca-64mjk\" (UID: \"b184c651-661a-4d84-bb0a-75107bf5de1b\") " pod="openshift-image-registry/node-ca-64mjk" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.673186 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t68cm\" (UniqueName: \"kubernetes.io/projected/b184c651-661a-4d84-bb0a-75107bf5de1b-kube-api-access-t68cm\") pod \"node-ca-64mjk\" (UID: \"b184c651-661a-4d84-bb0a-75107bf5de1b\") " pod="openshift-image-registry/node-ca-64mjk" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.673222 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b184c651-661a-4d84-bb0a-75107bf5de1b-host\") pod \"node-ca-64mjk\" (UID: \"b184c651-661a-4d84-bb0a-75107bf5de1b\") " pod="openshift-image-registry/node-ca-64mjk" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.678594 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.686196 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.686255 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.686271 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.686292 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.686305 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:59Z","lastTransitionTime":"2025-12-01T17:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.694894 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.707964 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.720299 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.737419 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.753663 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.767313 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.773956 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b184c651-661a-4d84-bb0a-75107bf5de1b-serviceca\") pod \"node-ca-64mjk\" (UID: \"b184c651-661a-4d84-bb0a-75107bf5de1b\") " pod="openshift-image-registry/node-ca-64mjk" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.774035 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t68cm\" (UniqueName: \"kubernetes.io/projected/b184c651-661a-4d84-bb0a-75107bf5de1b-kube-api-access-t68cm\") pod \"node-ca-64mjk\" (UID: \"b184c651-661a-4d84-bb0a-75107bf5de1b\") " pod="openshift-image-registry/node-ca-64mjk" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.774063 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b184c651-661a-4d84-bb0a-75107bf5de1b-host\") pod \"node-ca-64mjk\" (UID: \"b184c651-661a-4d84-bb0a-75107bf5de1b\") " pod="openshift-image-registry/node-ca-64mjk" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.774122 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b184c651-661a-4d84-bb0a-75107bf5de1b-host\") pod \"node-ca-64mjk\" (UID: \"b184c651-661a-4d84-bb0a-75107bf5de1b\") " pod="openshift-image-registry/node-ca-64mjk" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.776112 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b184c651-661a-4d84-bb0a-75107bf5de1b-serviceca\") pod \"node-ca-64mjk\" (UID: \"b184c651-661a-4d84-bb0a-75107bf5de1b\") " pod="openshift-image-registry/node-ca-64mjk" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.779257 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.813764 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.819185 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t68cm\" (UniqueName: \"kubernetes.io/projected/b184c651-661a-4d84-bb0a-75107bf5de1b-kube-api-access-t68cm\") pod \"node-ca-64mjk\" (UID: \"b184c651-661a-4d84-bb0a-75107bf5de1b\") " pod="openshift-image-registry/node-ca-64mjk" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.834842 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.834922 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.834976 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.835003 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.835021 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:59Z","lastTransitionTime":"2025-12-01T17:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.854264 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.864379 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.876699 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.889806 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.901360 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.916204 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:25:59Z is after 2025-08-24T17:21:41Z" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.939167 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.939217 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.939230 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.939250 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.939292 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:25:59Z","lastTransitionTime":"2025-12-01T17:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:25:59 crc kubenswrapper[4868]: I1201 17:25:59.999231 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-64mjk" Dec 01 17:26:00 crc kubenswrapper[4868]: W1201 17:26:00.019256 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb184c651_661a_4d84_bb0a_75107bf5de1b.slice/crio-1d1705293af2058a9fea4de04e232983d5296ac4cf93e4d5c39b0a53200a6682 WatchSource:0}: Error finding container 1d1705293af2058a9fea4de04e232983d5296ac4cf93e4d5c39b0a53200a6682: Status 404 returned error can't find the container with id 1d1705293af2058a9fea4de04e232983d5296ac4cf93e4d5c39b0a53200a6682 Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.047884 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.048385 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.048428 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.048456 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.048471 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:00Z","lastTransitionTime":"2025-12-01T17:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.156479 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.156555 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.156571 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.156595 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.156610 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:00Z","lastTransitionTime":"2025-12-01T17:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.172047 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:00 crc kubenswrapper[4868]: E1201 17:26:00.172296 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.261671 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.261710 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.261720 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.261736 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.261746 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:00Z","lastTransitionTime":"2025-12-01T17:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.372572 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.372624 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.372634 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.372650 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.372663 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:00Z","lastTransitionTime":"2025-12-01T17:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.423166 4868 generic.go:334] "Generic (PLEG): container finished" podID="46acee96-a86d-4abf-a79f-7cfc61fbab92" containerID="144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21" exitCode=0 Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.423231 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" event={"ID":"46acee96-a86d-4abf-a79f-7cfc61fbab92","Type":"ContainerDied","Data":"144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21"} Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.425157 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-64mjk" event={"ID":"b184c651-661a-4d84-bb0a-75107bf5de1b","Type":"ContainerStarted","Data":"4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7"} Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.425206 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-64mjk" event={"ID":"b184c651-661a-4d84-bb0a-75107bf5de1b","Type":"ContainerStarted","Data":"1d1705293af2058a9fea4de04e232983d5296ac4cf93e4d5c39b0a53200a6682"} Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.443723 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.458864 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.478709 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.478755 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.478769 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.478788 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.478798 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:00Z","lastTransitionTime":"2025-12-01T17:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.482502 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.495067 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.510624 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.544493 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.559047 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.579078 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.585698 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.585760 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.585777 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.585800 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.585817 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:00Z","lastTransitionTime":"2025-12-01T17:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.596683 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.613321 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.627978 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.642764 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.657619 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.672556 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.688339 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.689031 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.689073 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.689087 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.689105 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.689118 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:00Z","lastTransitionTime":"2025-12-01T17:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.702780 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.717157 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.730579 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.744628 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.763284 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.772967 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.784910 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.790754 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.790775 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.790783 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.790799 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.790809 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:00Z","lastTransitionTime":"2025-12-01T17:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.797906 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.810545 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.824244 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.837732 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.848492 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.865708 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:00Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.894003 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.894049 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.894057 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.894073 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.894084 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:00Z","lastTransitionTime":"2025-12-01T17:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.996661 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.996710 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.996724 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.996747 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:00 crc kubenswrapper[4868]: I1201 17:26:00.996762 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:00Z","lastTransitionTime":"2025-12-01T17:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.099329 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.099380 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.099395 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.099415 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.099428 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:01Z","lastTransitionTime":"2025-12-01T17:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.170917 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.171140 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:01 crc kubenswrapper[4868]: E1201 17:26:01.171234 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:01 crc kubenswrapper[4868]: E1201 17:26:01.171378 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.202007 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.202048 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.202057 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.202074 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.202085 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:01Z","lastTransitionTime":"2025-12-01T17:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.305047 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.305095 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.305115 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.305164 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.305181 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:01Z","lastTransitionTime":"2025-12-01T17:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.408649 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.408693 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.408702 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.408717 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.408728 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:01Z","lastTransitionTime":"2025-12-01T17:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.432978 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerStarted","Data":"1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0"} Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.435806 4868 generic.go:334] "Generic (PLEG): container finished" podID="46acee96-a86d-4abf-a79f-7cfc61fbab92" containerID="176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab" exitCode=0 Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.435857 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" event={"ID":"46acee96-a86d-4abf-a79f-7cfc61fbab92","Type":"ContainerDied","Data":"176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab"} Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.455379 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.467769 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.481085 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.494906 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.507297 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.511054 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.511079 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.511088 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.511106 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.511115 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:01Z","lastTransitionTime":"2025-12-01T17:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.521374 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.533477 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.545278 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.564876 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.580462 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.599305 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.611540 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.613312 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.613340 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.613351 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.613368 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.613379 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:01Z","lastTransitionTime":"2025-12-01T17:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.626240 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.635311 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.635379 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.635401 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.635428 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.635448 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:01Z","lastTransitionTime":"2025-12-01T17:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.641668 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: E1201 17:26:01.651037 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.655581 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.655629 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.655656 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.655682 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.655696 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:01Z","lastTransitionTime":"2025-12-01T17:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:01 crc kubenswrapper[4868]: E1201 17:26:01.674377 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.679371 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.679403 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.679413 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.679430 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.679440 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:01Z","lastTransitionTime":"2025-12-01T17:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:01 crc kubenswrapper[4868]: E1201 17:26:01.692648 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.696444 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.696491 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.696508 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.696527 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.696539 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:01Z","lastTransitionTime":"2025-12-01T17:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:01 crc kubenswrapper[4868]: E1201 17:26:01.709691 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.714837 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.714871 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.714881 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.714898 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.714908 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:01Z","lastTransitionTime":"2025-12-01T17:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:01 crc kubenswrapper[4868]: E1201 17:26:01.735925 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:01Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:01 crc kubenswrapper[4868]: E1201 17:26:01.736088 4868 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.738014 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.738047 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.738058 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.738076 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.738087 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:01Z","lastTransitionTime":"2025-12-01T17:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.840470 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.840511 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.840522 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.840539 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.840549 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:01Z","lastTransitionTime":"2025-12-01T17:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.942619 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.942662 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.942674 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.942693 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:01 crc kubenswrapper[4868]: I1201 17:26:01.942705 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:01Z","lastTransitionTime":"2025-12-01T17:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.045441 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.045490 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.045502 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.045520 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.045533 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:02Z","lastTransitionTime":"2025-12-01T17:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.147741 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.147776 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.147784 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.147800 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.147810 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:02Z","lastTransitionTime":"2025-12-01T17:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.171325 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:02 crc kubenswrapper[4868]: E1201 17:26:02.171491 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.249854 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.249925 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.249975 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.250007 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.250027 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:02Z","lastTransitionTime":"2025-12-01T17:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.353060 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.353122 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.353140 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.353166 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.353187 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:02Z","lastTransitionTime":"2025-12-01T17:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.445417 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" event={"ID":"46acee96-a86d-4abf-a79f-7cfc61fbab92","Type":"ContainerStarted","Data":"385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea"} Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.454992 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.455033 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.455047 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.455092 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.455106 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:02Z","lastTransitionTime":"2025-12-01T17:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.463548 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:02Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.475628 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:02Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.493841 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:02Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.506652 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:02Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.522013 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:02Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.543343 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:02Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.557800 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.557849 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.557859 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.557877 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.557889 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:02Z","lastTransitionTime":"2025-12-01T17:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.560173 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:02Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.584306 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:02Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.593767 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:02Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.609470 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:02Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.654839 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:02Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.659664 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.659701 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.659712 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.659728 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.659737 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:02Z","lastTransitionTime":"2025-12-01T17:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.675242 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:02Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.704709 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:02Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.724936 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:02Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.762568 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.762617 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.762630 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.762652 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.762665 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:02Z","lastTransitionTime":"2025-12-01T17:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.864740 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.864770 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.864780 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.864795 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.864805 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:02Z","lastTransitionTime":"2025-12-01T17:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.922635 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.922733 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:02 crc kubenswrapper[4868]: E1201 17:26:02.922833 4868 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:26:02 crc kubenswrapper[4868]: E1201 17:26:02.922838 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:26:18.922804481 +0000 UTC m=+51.293914892 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:26:02 crc kubenswrapper[4868]: E1201 17:26:02.922957 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:26:18.922929114 +0000 UTC m=+51.294039525 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.923023 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:02 crc kubenswrapper[4868]: E1201 17:26:02.923272 4868 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:26:02 crc kubenswrapper[4868]: E1201 17:26:02.923427 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:26:18.923376117 +0000 UTC m=+51.294486528 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.966961 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.967011 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.967024 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.967043 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:02 crc kubenswrapper[4868]: I1201 17:26:02.967055 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:02Z","lastTransitionTime":"2025-12-01T17:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.024433 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.024518 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:03 crc kubenswrapper[4868]: E1201 17:26:03.024688 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:26:03 crc kubenswrapper[4868]: E1201 17:26:03.024738 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:26:03 crc kubenswrapper[4868]: E1201 17:26:03.024752 4868 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:26:03 crc kubenswrapper[4868]: E1201 17:26:03.024816 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 17:26:19.024794475 +0000 UTC m=+51.395904886 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:26:03 crc kubenswrapper[4868]: E1201 17:26:03.024705 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:26:03 crc kubenswrapper[4868]: E1201 17:26:03.024849 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:26:03 crc kubenswrapper[4868]: E1201 17:26:03.024867 4868 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:26:03 crc kubenswrapper[4868]: E1201 17:26:03.024958 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 17:26:19.024921389 +0000 UTC m=+51.396031820 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.069210 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.069247 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.069257 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.069273 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.069282 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:03Z","lastTransitionTime":"2025-12-01T17:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.171058 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.171103 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:03 crc kubenswrapper[4868]: E1201 17:26:03.171218 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:03 crc kubenswrapper[4868]: E1201 17:26:03.171391 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.172174 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.172253 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.172262 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.172280 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.172292 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:03Z","lastTransitionTime":"2025-12-01T17:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.275811 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.275857 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.275867 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.275884 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.275895 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:03Z","lastTransitionTime":"2025-12-01T17:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.379554 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.379646 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.379674 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.379724 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.379752 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:03Z","lastTransitionTime":"2025-12-01T17:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.454193 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerStarted","Data":"213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32"} Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.454712 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.454735 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.471989 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.482673 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.482718 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.482730 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.482747 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.482758 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:03Z","lastTransitionTime":"2025-12-01T17:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.489820 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.505398 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.525120 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.544609 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.563150 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.582696 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.584909 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.584982 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.584999 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.585022 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.585037 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:03Z","lastTransitionTime":"2025-12-01T17:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.600746 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.620877 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.632132 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.639828 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.661283 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.683615 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.687904 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.688013 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.688032 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.688059 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.688079 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:03Z","lastTransitionTime":"2025-12-01T17:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.694268 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.694676 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.708528 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.722971 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.739206 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.752426 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.763803 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.781055 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.791828 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.791863 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.791874 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.791891 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.791904 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:03Z","lastTransitionTime":"2025-12-01T17:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.803308 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.814650 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.826866 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.841608 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.855352 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.875133 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.890443 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.894445 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.894506 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.894522 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.894546 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.894561 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:03Z","lastTransitionTime":"2025-12-01T17:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.904510 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.916844 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.930865 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.997867 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.997926 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.997969 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.998005 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:03 crc kubenswrapper[4868]: I1201 17:26:03.998019 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:03Z","lastTransitionTime":"2025-12-01T17:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.102198 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.102609 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.102627 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.102651 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.102669 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:04Z","lastTransitionTime":"2025-12-01T17:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.171967 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:04 crc kubenswrapper[4868]: E1201 17:26:04.172150 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.205059 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.205103 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.205118 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.205138 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.205152 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:04Z","lastTransitionTime":"2025-12-01T17:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.308042 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.308121 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.308141 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.308172 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.308192 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:04Z","lastTransitionTime":"2025-12-01T17:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.411099 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.411156 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.411169 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.411194 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.411206 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:04Z","lastTransitionTime":"2025-12-01T17:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.461332 4868 generic.go:334] "Generic (PLEG): container finished" podID="46acee96-a86d-4abf-a79f-7cfc61fbab92" containerID="385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea" exitCode=0 Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.461406 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" event={"ID":"46acee96-a86d-4abf-a79f-7cfc61fbab92","Type":"ContainerDied","Data":"385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea"} Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.461799 4868 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.485453 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:04Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.503817 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:04Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.513867 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.513911 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.513923 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.513955 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.513968 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:04Z","lastTransitionTime":"2025-12-01T17:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.519933 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:04Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.537126 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:04Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.555765 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:04Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.569732 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:04Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.589976 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:04Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.612046 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:04Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.618058 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.618099 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.618110 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.618127 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.618138 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:04Z","lastTransitionTime":"2025-12-01T17:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.625833 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:04Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.640096 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:04Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.657869 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:04Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.672116 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:04Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.685700 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:04Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.700407 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:04Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.721369 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.721423 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.721433 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.721452 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.721463 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:04Z","lastTransitionTime":"2025-12-01T17:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.824501 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.824546 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.824557 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.824576 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.824588 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:04Z","lastTransitionTime":"2025-12-01T17:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.927275 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.927356 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.927370 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.927395 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:04 crc kubenswrapper[4868]: I1201 17:26:04.927411 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:04Z","lastTransitionTime":"2025-12-01T17:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.030381 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.030449 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.030467 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.030494 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.030512 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:05Z","lastTransitionTime":"2025-12-01T17:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.133884 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.133937 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.133960 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.133980 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.133991 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:05Z","lastTransitionTime":"2025-12-01T17:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.172118 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.172120 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:05 crc kubenswrapper[4868]: E1201 17:26:05.172338 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:05 crc kubenswrapper[4868]: E1201 17:26:05.172447 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.241610 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.241700 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.241739 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.241762 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.241776 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:05Z","lastTransitionTime":"2025-12-01T17:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.344967 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.345011 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.345023 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.345045 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.345057 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:05Z","lastTransitionTime":"2025-12-01T17:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.448747 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.448795 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.448803 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.448819 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.448830 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:05Z","lastTransitionTime":"2025-12-01T17:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.468908 4868 generic.go:334] "Generic (PLEG): container finished" podID="46acee96-a86d-4abf-a79f-7cfc61fbab92" containerID="9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886" exitCode=0 Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.468970 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" event={"ID":"46acee96-a86d-4abf-a79f-7cfc61fbab92","Type":"ContainerDied","Data":"9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886"} Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.469114 4868 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.484553 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:05Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.507514 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:05Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.525847 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:05Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.540967 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:05Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.551792 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.551823 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.551834 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.551852 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.551865 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:05Z","lastTransitionTime":"2025-12-01T17:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.556190 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:05Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.576121 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:05Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.605001 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:05Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.618369 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:05Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.634892 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:05Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.656184 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.656230 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.656242 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.656264 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.656277 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:05Z","lastTransitionTime":"2025-12-01T17:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.656757 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:05Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.670848 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:05Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.684293 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:05Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.723388 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:05Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.741120 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:05Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.761188 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.761232 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.761244 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.761260 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.761271 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:05Z","lastTransitionTime":"2025-12-01T17:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.862932 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.862997 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.863006 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.863020 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.863033 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:05Z","lastTransitionTime":"2025-12-01T17:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.965722 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.965760 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.965768 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.965785 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:05 crc kubenswrapper[4868]: I1201 17:26:05.965794 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:05Z","lastTransitionTime":"2025-12-01T17:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.068307 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.068344 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.068353 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.068371 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.068381 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:06Z","lastTransitionTime":"2025-12-01T17:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.171047 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:06 crc kubenswrapper[4868]: E1201 17:26:06.171197 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.171299 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.171330 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.171342 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.171358 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.171371 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:06Z","lastTransitionTime":"2025-12-01T17:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.273339 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.273382 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.273393 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.273409 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.273417 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:06Z","lastTransitionTime":"2025-12-01T17:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.375502 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.375540 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.375549 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.375566 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.375586 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:06Z","lastTransitionTime":"2025-12-01T17:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.480322 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.480366 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.480375 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.480389 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.480399 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:06Z","lastTransitionTime":"2025-12-01T17:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.498470 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" event={"ID":"46acee96-a86d-4abf-a79f-7cfc61fbab92","Type":"ContainerStarted","Data":"cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4"} Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.514656 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:06Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.529524 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:06Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.544350 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:06Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.559058 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:06Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.572086 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:06Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.582675 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.582712 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.582722 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.582740 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.582750 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:06Z","lastTransitionTime":"2025-12-01T17:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.590107 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:06Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.609347 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:06Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.622964 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:06Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.642741 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:06Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.657531 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:06Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.683813 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:06Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.685394 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.685499 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.685718 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.685899 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.686099 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:06Z","lastTransitionTime":"2025-12-01T17:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.700853 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:06Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.716419 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:06Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.730387 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:06Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.788768 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.788813 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.788827 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.788847 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.788859 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:06Z","lastTransitionTime":"2025-12-01T17:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.909574 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.909800 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.909867 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.909978 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:06 crc kubenswrapper[4868]: I1201 17:26:06.910049 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:06Z","lastTransitionTime":"2025-12-01T17:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.101999 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.102330 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.102344 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.102373 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.102391 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:07Z","lastTransitionTime":"2025-12-01T17:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.171144 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.171224 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:07 crc kubenswrapper[4868]: E1201 17:26:07.171572 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:07 crc kubenswrapper[4868]: E1201 17:26:07.171614 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.180709 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72"] Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.181474 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" Dec 01 17:26:07 crc kubenswrapper[4868]: W1201 17:26:07.184223 4868 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert": failed to list *v1.Secret: secrets "ovn-control-plane-metrics-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 01 17:26:07 crc kubenswrapper[4868]: E1201 17:26:07.184510 4868 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-control-plane-metrics-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-control-plane-metrics-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 17:26:07 crc kubenswrapper[4868]: W1201 17:26:07.184549 4868 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd": failed to list *v1.Secret: secrets "ovn-kubernetes-control-plane-dockercfg-gs7dd" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 01 17:26:07 crc kubenswrapper[4868]: E1201 17:26:07.184762 4868 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-control-plane-dockercfg-gs7dd\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-control-plane-dockercfg-gs7dd\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.199651 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:07Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.204343 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.204375 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.204386 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.204404 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.204416 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:07Z","lastTransitionTime":"2025-12-01T17:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.212107 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:07Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.222621 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:07Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.235340 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:07Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.248221 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:07Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.267761 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:07Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.277329 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:07Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.283070 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c7e1f752-fd18-475d-af3b-8af3a51bc2d5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-c7p72\" (UID: \"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.283132 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c7e1f752-fd18-475d-af3b-8af3a51bc2d5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-c7p72\" (UID: \"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.283154 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdnmj\" (UniqueName: \"kubernetes.io/projected/c7e1f752-fd18-475d-af3b-8af3a51bc2d5-kube-api-access-bdnmj\") pod \"ovnkube-control-plane-749d76644c-c7p72\" (UID: \"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.283193 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c7e1f752-fd18-475d-af3b-8af3a51bc2d5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-c7p72\" (UID: \"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.290362 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:07Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.306006 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:07Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.307590 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.307721 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.307837 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.307919 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.308024 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:07Z","lastTransitionTime":"2025-12-01T17:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.319233 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:07Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.334321 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:07Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.347126 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:07Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.361067 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:07Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.374909 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:07Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.383617 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdnmj\" (UniqueName: \"kubernetes.io/projected/c7e1f752-fd18-475d-af3b-8af3a51bc2d5-kube-api-access-bdnmj\") pod \"ovnkube-control-plane-749d76644c-c7p72\" (UID: \"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.383808 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c7e1f752-fd18-475d-af3b-8af3a51bc2d5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-c7p72\" (UID: \"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.383932 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c7e1f752-fd18-475d-af3b-8af3a51bc2d5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-c7p72\" (UID: \"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.384243 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c7e1f752-fd18-475d-af3b-8af3a51bc2d5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-c7p72\" (UID: \"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.384793 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c7e1f752-fd18-475d-af3b-8af3a51bc2d5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-c7p72\" (UID: \"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.384971 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c7e1f752-fd18-475d-af3b-8af3a51bc2d5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-c7p72\" (UID: \"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.387676 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:07Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.402617 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdnmj\" (UniqueName: \"kubernetes.io/projected/c7e1f752-fd18-475d-af3b-8af3a51bc2d5-kube-api-access-bdnmj\") pod \"ovnkube-control-plane-749d76644c-c7p72\" (UID: \"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.410968 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.411009 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.411018 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.411035 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.411046 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:07Z","lastTransitionTime":"2025-12-01T17:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.513141 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.513179 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.513192 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.513210 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.513221 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:07Z","lastTransitionTime":"2025-12-01T17:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.615606 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.616309 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.616423 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.616522 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.616609 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:07Z","lastTransitionTime":"2025-12-01T17:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.719264 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.719300 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.719309 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.719326 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.719336 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:07Z","lastTransitionTime":"2025-12-01T17:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.822185 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.822467 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.822588 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.822698 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.822785 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:07Z","lastTransitionTime":"2025-12-01T17:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.925524 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.925581 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.925594 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.925615 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:07 crc kubenswrapper[4868]: I1201 17:26:07.925628 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:07Z","lastTransitionTime":"2025-12-01T17:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.028365 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.028407 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.028420 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.028437 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.028446 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:08Z","lastTransitionTime":"2025-12-01T17:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.036069 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.131045 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.131104 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.131116 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.131136 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.131149 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:08Z","lastTransitionTime":"2025-12-01T17:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.172155 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:08 crc kubenswrapper[4868]: E1201 17:26:08.172287 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.193816 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.208489 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.220045 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.233057 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.233762 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.233806 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.233820 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.233838 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.233850 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:08Z","lastTransitionTime":"2025-12-01T17:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.245364 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.259542 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.273051 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.283436 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.300007 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.336169 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.336255 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.336274 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.336302 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.336319 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:08Z","lastTransitionTime":"2025-12-01T17:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.338280 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.363414 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: E1201 17:26:08.386079 4868 secret.go:188] Couldn't get secret openshift-ovn-kubernetes/ovn-control-plane-metrics-cert: failed to sync secret cache: timed out waiting for the condition Dec 01 17:26:08 crc kubenswrapper[4868]: E1201 17:26:08.386272 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c7e1f752-fd18-475d-af3b-8af3a51bc2d5-ovn-control-plane-metrics-cert podName:c7e1f752-fd18-475d-af3b-8af3a51bc2d5 nodeName:}" failed. No retries permitted until 2025-12-01 17:26:08.886241324 +0000 UTC m=+41.257351735 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ovn-control-plane-metrics-cert" (UniqueName: "kubernetes.io/secret/c7e1f752-fd18-475d-af3b-8af3a51bc2d5-ovn-control-plane-metrics-cert") pod "ovnkube-control-plane-749d76644c-c7p72" (UID: "c7e1f752-fd18-475d-af3b-8af3a51bc2d5") : failed to sync secret cache: timed out waiting for the condition Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.395142 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: E1201 17:26:08.408985 4868 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5baca705_90cc_4198_b641_214a2b915086.slice/crio-conmon-213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32.scope\": RecentStats: unable to find data in memory cache]" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.418569 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.433014 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.438257 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.438286 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.438295 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.438312 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.438322 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:08Z","lastTransitionTime":"2025-12-01T17:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.448987 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.483026 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.507493 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovnkube-controller/0.log" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.511069 4868 generic.go:334] "Generic (PLEG): container finished" podID="5baca705-90cc-4198-b641-214a2b915086" containerID="213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32" exitCode=1 Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.511135 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerDied","Data":"213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32"} Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.511996 4868 scope.go:117] "RemoveContainer" containerID="213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.527795 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.541544 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.541593 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.541605 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.541625 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.541637 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:08Z","lastTransitionTime":"2025-12-01T17:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.544311 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.557099 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.570342 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.582232 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.594156 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.605215 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.619754 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.629533 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.640127 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.643813 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.643865 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.643874 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.643893 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.643907 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:08Z","lastTransitionTime":"2025-12-01T17:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.651483 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.660742 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.674894 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.692563 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"tory.go:141\\\\nI1201 17:26:07.251211 6159 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:07.251238 6159 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:07.251282 6159 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.251301 6159 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.251394 6159 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.253796 6159 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 17:26:07.253814 6159 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 17:26:07.253846 6159 factory.go:656] Stopping watch factory\\\\nI1201 17:26:07.253852 6159 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 17:26:07.253870 6159 ovnkube.go:599] Stopped ovnkube\\\\nI1201 17\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.702462 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.704405 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-nxfkx"] Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.705013 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:08 crc kubenswrapper[4868]: E1201 17:26:08.705090 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.721466 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.735008 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.745044 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.745882 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.745910 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.745919 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.745934 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.745965 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:08Z","lastTransitionTime":"2025-12-01T17:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.759102 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.782316 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"tory.go:141\\\\nI1201 17:26:07.251211 6159 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:07.251238 6159 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:07.251282 6159 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.251301 6159 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.251394 6159 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.253796 6159 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 17:26:07.253814 6159 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 17:26:07.253846 6159 factory.go:656] Stopping watch factory\\\\nI1201 17:26:07.253852 6159 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 17:26:07.253870 6159 ovnkube.go:599] Stopped ovnkube\\\\nI1201 17\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.791495 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.800070 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs\") pod \"network-metrics-daemon-nxfkx\" (UID: \"4afd9a0c-e8c7-42ec-9771-036c3441cbce\") " pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.800346 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zc4h\" (UniqueName: \"kubernetes.io/projected/4afd9a0c-e8c7-42ec-9771-036c3441cbce-kube-api-access-5zc4h\") pod \"network-metrics-daemon-nxfkx\" (UID: \"4afd9a0c-e8c7-42ec-9771-036c3441cbce\") " pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.800546 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.811707 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.823705 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.837092 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.848568 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.848607 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.848618 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.848636 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.848645 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:08Z","lastTransitionTime":"2025-12-01T17:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.852639 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.870093 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.884253 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.894636 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.901116 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs\") pod \"network-metrics-daemon-nxfkx\" (UID: \"4afd9a0c-e8c7-42ec-9771-036c3441cbce\") " pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.901161 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c7e1f752-fd18-475d-af3b-8af3a51bc2d5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-c7p72\" (UID: \"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.901201 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zc4h\" (UniqueName: \"kubernetes.io/projected/4afd9a0c-e8c7-42ec-9771-036c3441cbce-kube-api-access-5zc4h\") pod \"network-metrics-daemon-nxfkx\" (UID: \"4afd9a0c-e8c7-42ec-9771-036c3441cbce\") " pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:08 crc kubenswrapper[4868]: E1201 17:26:08.901268 4868 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:26:08 crc kubenswrapper[4868]: E1201 17:26:08.901328 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs podName:4afd9a0c-e8c7-42ec-9771-036c3441cbce nodeName:}" failed. No retries permitted until 2025-12-01 17:26:09.40131295 +0000 UTC m=+41.772423361 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs") pod "network-metrics-daemon-nxfkx" (UID: "4afd9a0c-e8c7-42ec-9771-036c3441cbce") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.904862 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c7e1f752-fd18-475d-af3b-8af3a51bc2d5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-c7p72\" (UID: \"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.907679 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.918617 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zc4h\" (UniqueName: \"kubernetes.io/projected/4afd9a0c-e8c7-42ec-9771-036c3441cbce-kube-api-access-5zc4h\") pod \"network-metrics-daemon-nxfkx\" (UID: \"4afd9a0c-e8c7-42ec-9771-036c3441cbce\") " pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.920132 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.952727 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.952767 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.952783 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.952801 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.952812 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:08Z","lastTransitionTime":"2025-12-01T17:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:08 crc kubenswrapper[4868]: I1201 17:26:08.996234 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.055815 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.056067 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.056076 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.056092 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.056103 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:09Z","lastTransitionTime":"2025-12-01T17:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.158506 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.158543 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.158553 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.158576 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.158588 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:09Z","lastTransitionTime":"2025-12-01T17:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.171569 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:09 crc kubenswrapper[4868]: E1201 17:26:09.171696 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.171778 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:09 crc kubenswrapper[4868]: E1201 17:26:09.171841 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.261675 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.261721 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.261730 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.261744 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.261753 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:09Z","lastTransitionTime":"2025-12-01T17:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.364389 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.364432 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.364443 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.364461 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.364471 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:09Z","lastTransitionTime":"2025-12-01T17:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.405701 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs\") pod \"network-metrics-daemon-nxfkx\" (UID: \"4afd9a0c-e8c7-42ec-9771-036c3441cbce\") " pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:09 crc kubenswrapper[4868]: E1201 17:26:09.405917 4868 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:26:09 crc kubenswrapper[4868]: E1201 17:26:09.406054 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs podName:4afd9a0c-e8c7-42ec-9771-036c3441cbce nodeName:}" failed. No retries permitted until 2025-12-01 17:26:10.406025807 +0000 UTC m=+42.777136218 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs") pod "network-metrics-daemon-nxfkx" (UID: "4afd9a0c-e8c7-42ec-9771-036c3441cbce") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.467063 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.467103 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.467113 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.467129 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.467137 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:09Z","lastTransitionTime":"2025-12-01T17:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.516143 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovnkube-controller/0.log" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.519169 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerStarted","Data":"48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee"} Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.519584 4868 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.520472 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" event={"ID":"c7e1f752-fd18-475d-af3b-8af3a51bc2d5","Type":"ContainerStarted","Data":"b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df"} Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.520525 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" event={"ID":"c7e1f752-fd18-475d-af3b-8af3a51bc2d5","Type":"ContainerStarted","Data":"a53e8754a3527566da63a777c8322d50543b529c1eb536a222b0aea253cc999b"} Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.541251 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.555363 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.569974 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.570011 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.570023 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.570044 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.570057 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:09Z","lastTransitionTime":"2025-12-01T17:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.571075 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.584923 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.596936 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.614046 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.634803 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"tory.go:141\\\\nI1201 17:26:07.251211 6159 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:07.251238 6159 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:07.251282 6159 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.251301 6159 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.251394 6159 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.253796 6159 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 17:26:07.253814 6159 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 17:26:07.253846 6159 factory.go:656] Stopping watch factory\\\\nI1201 17:26:07.253852 6159 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 17:26:07.253870 6159 ovnkube.go:599] Stopped ovnkube\\\\nI1201 17\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.647587 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.661224 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.672294 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.672323 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.672331 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.672346 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.672354 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:09Z","lastTransitionTime":"2025-12-01T17:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.677278 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.691965 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.709498 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.725599 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.736548 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.748138 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.760479 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:09Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.778177 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.778233 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.778244 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.778264 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.778277 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:09Z","lastTransitionTime":"2025-12-01T17:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.881042 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.881111 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.881128 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.881152 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.881170 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:09Z","lastTransitionTime":"2025-12-01T17:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.984903 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.984972 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.984982 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.985000 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:09 crc kubenswrapper[4868]: I1201 17:26:09.985010 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:09Z","lastTransitionTime":"2025-12-01T17:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.088217 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.088284 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.088302 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.088327 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.088347 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:10Z","lastTransitionTime":"2025-12-01T17:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.171436 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:10 crc kubenswrapper[4868]: E1201 17:26:10.171669 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.172099 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:10 crc kubenswrapper[4868]: E1201 17:26:10.172269 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.191683 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.191735 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.191749 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.191769 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.191781 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:10Z","lastTransitionTime":"2025-12-01T17:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.294493 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.294534 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.294546 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.294562 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.294573 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:10Z","lastTransitionTime":"2025-12-01T17:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.398342 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.398393 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.398406 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.398427 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.398439 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:10Z","lastTransitionTime":"2025-12-01T17:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.418288 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs\") pod \"network-metrics-daemon-nxfkx\" (UID: \"4afd9a0c-e8c7-42ec-9771-036c3441cbce\") " pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:10 crc kubenswrapper[4868]: E1201 17:26:10.418476 4868 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:26:10 crc kubenswrapper[4868]: E1201 17:26:10.418593 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs podName:4afd9a0c-e8c7-42ec-9771-036c3441cbce nodeName:}" failed. No retries permitted until 2025-12-01 17:26:12.418559279 +0000 UTC m=+44.789669700 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs") pod "network-metrics-daemon-nxfkx" (UID: "4afd9a0c-e8c7-42ec-9771-036c3441cbce") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.501130 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.501178 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.501194 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.501215 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.501229 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:10Z","lastTransitionTime":"2025-12-01T17:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.526222 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovnkube-controller/1.log" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.527091 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovnkube-controller/0.log" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.530355 4868 generic.go:334] "Generic (PLEG): container finished" podID="5baca705-90cc-4198-b641-214a2b915086" containerID="48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee" exitCode=1 Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.530410 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerDied","Data":"48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee"} Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.530499 4868 scope.go:117] "RemoveContainer" containerID="213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.531426 4868 scope.go:117] "RemoveContainer" containerID="48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee" Dec 01 17:26:10 crc kubenswrapper[4868]: E1201 17:26:10.531602 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-q9z6l_openshift-ovn-kubernetes(5baca705-90cc-4198-b641-214a2b915086)\"" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podUID="5baca705-90cc-4198-b641-214a2b915086" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.532864 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" event={"ID":"c7e1f752-fd18-475d-af3b-8af3a51bc2d5","Type":"ContainerStarted","Data":"1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8"} Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.547577 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.563468 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.578764 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.594255 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.604272 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.604333 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.604346 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.604369 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.604385 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:10Z","lastTransitionTime":"2025-12-01T17:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.607643 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.623656 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.645130 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"tory.go:141\\\\nI1201 17:26:07.251211 6159 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:07.251238 6159 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:07.251282 6159 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.251301 6159 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.251394 6159 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.253796 6159 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 17:26:07.253814 6159 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 17:26:07.253846 6159 factory.go:656] Stopping watch factory\\\\nI1201 17:26:07.253852 6159 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 17:26:07.253870 6159 ovnkube.go:599] Stopped ovnkube\\\\nI1201 17\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:09Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.556807 6333 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.556910 6333 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.557032 6333 factory.go:656] Stopping watch factory\\\\nI1201 17:26:09.557052 6333 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557341 6333 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557758 6333 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:09.559974 6333 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1201 17:26:09.559996 6333 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1201 17:26:09.560056 6333 ovnkube.go:599] Stopped ovnkube\\\\nI1201 17:26:09.560088 6333 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1201 17:26:09.560163 6333 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.659442 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.675070 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.689900 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.706622 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.708408 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.708460 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.708477 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.708496 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.708554 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:10Z","lastTransitionTime":"2025-12-01T17:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.723613 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.738823 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.755458 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.769000 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.781485 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.796633 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.807100 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.812724 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.812793 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.812812 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.812837 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.812860 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:10Z","lastTransitionTime":"2025-12-01T17:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.825269 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.843363 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.861774 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.884202 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"tory.go:141\\\\nI1201 17:26:07.251211 6159 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:07.251238 6159 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:07.251282 6159 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.251301 6159 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.251394 6159 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.253796 6159 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 17:26:07.253814 6159 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 17:26:07.253846 6159 factory.go:656] Stopping watch factory\\\\nI1201 17:26:07.253852 6159 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 17:26:07.253870 6159 ovnkube.go:599] Stopped ovnkube\\\\nI1201 17\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:09Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.556807 6333 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.556910 6333 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.557032 6333 factory.go:656] Stopping watch factory\\\\nI1201 17:26:09.557052 6333 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557341 6333 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557758 6333 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:09.559974 6333 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1201 17:26:09.559996 6333 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1201 17:26:09.560056 6333 ovnkube.go:599] Stopped ovnkube\\\\nI1201 17:26:09.560088 6333 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1201 17:26:09.560163 6333 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.901554 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.915348 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.916377 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.916523 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.916827 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.917052 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.917229 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:10Z","lastTransitionTime":"2025-12-01T17:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.929986 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.946901 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.962608 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.976123 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:10 crc kubenswrapper[4868]: I1201 17:26:10.989867 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:10Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.004193 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:11Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.015805 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:11Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.021515 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.021563 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.021574 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.021593 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.021605 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:11Z","lastTransitionTime":"2025-12-01T17:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.028025 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:11Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.124324 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.124355 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.124363 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.124380 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.124391 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:11Z","lastTransitionTime":"2025-12-01T17:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.171995 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:11 crc kubenswrapper[4868]: E1201 17:26:11.172198 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.172391 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:11 crc kubenswrapper[4868]: E1201 17:26:11.172568 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.227705 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.228030 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.228137 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.228230 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.228309 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:11Z","lastTransitionTime":"2025-12-01T17:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.330885 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.330927 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.330936 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.330980 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.330989 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:11Z","lastTransitionTime":"2025-12-01T17:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.433976 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.434015 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.434025 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.434043 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.434054 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:11Z","lastTransitionTime":"2025-12-01T17:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.535983 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.536016 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.536026 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.536063 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.536075 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:11Z","lastTransitionTime":"2025-12-01T17:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.538881 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovnkube-controller/1.log" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.638550 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.638606 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.638623 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.638644 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.638655 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:11Z","lastTransitionTime":"2025-12-01T17:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.740409 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.741075 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.741108 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.741135 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.741147 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:11Z","lastTransitionTime":"2025-12-01T17:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.843636 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.843677 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.843693 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.843710 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.843721 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:11Z","lastTransitionTime":"2025-12-01T17:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.946596 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.946647 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.946659 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.946678 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:11 crc kubenswrapper[4868]: I1201 17:26:11.946688 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:11Z","lastTransitionTime":"2025-12-01T17:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.049169 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.049232 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.049254 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.049284 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.049306 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:12Z","lastTransitionTime":"2025-12-01T17:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.092412 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.092454 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.092465 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.092484 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.092495 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:12Z","lastTransitionTime":"2025-12-01T17:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:12 crc kubenswrapper[4868]: E1201 17:26:12.107367 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:12Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.112729 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.112762 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.112771 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.112787 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.112796 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:12Z","lastTransitionTime":"2025-12-01T17:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:12 crc kubenswrapper[4868]: E1201 17:26:12.129352 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:12Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.133984 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.134042 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.134054 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.134073 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.134086 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:12Z","lastTransitionTime":"2025-12-01T17:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:12 crc kubenswrapper[4868]: E1201 17:26:12.147614 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:12Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.151291 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.151334 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.151348 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.151368 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.151382 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:12Z","lastTransitionTime":"2025-12-01T17:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:12 crc kubenswrapper[4868]: E1201 17:26:12.163836 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:12Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.167697 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.167746 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.167756 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.167776 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.167788 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:12Z","lastTransitionTime":"2025-12-01T17:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.171585 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.171596 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:12 crc kubenswrapper[4868]: E1201 17:26:12.171740 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:12 crc kubenswrapper[4868]: E1201 17:26:12.171781 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:12 crc kubenswrapper[4868]: E1201 17:26:12.184187 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:12Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:12 crc kubenswrapper[4868]: E1201 17:26:12.184340 4868 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.186073 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.186100 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.186108 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.186125 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.186139 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:12Z","lastTransitionTime":"2025-12-01T17:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.289751 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.289792 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.289800 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.289822 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.289835 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:12Z","lastTransitionTime":"2025-12-01T17:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.400828 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.400879 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.400891 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.400911 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.400922 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:12Z","lastTransitionTime":"2025-12-01T17:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.441640 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs\") pod \"network-metrics-daemon-nxfkx\" (UID: \"4afd9a0c-e8c7-42ec-9771-036c3441cbce\") " pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:12 crc kubenswrapper[4868]: E1201 17:26:12.441842 4868 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:26:12 crc kubenswrapper[4868]: E1201 17:26:12.442013 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs podName:4afd9a0c-e8c7-42ec-9771-036c3441cbce nodeName:}" failed. No retries permitted until 2025-12-01 17:26:16.441920083 +0000 UTC m=+48.813030484 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs") pod "network-metrics-daemon-nxfkx" (UID: "4afd9a0c-e8c7-42ec-9771-036c3441cbce") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.503479 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.503563 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.503595 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.503628 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.503654 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:12Z","lastTransitionTime":"2025-12-01T17:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.606375 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.606685 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.606764 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.606840 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.606924 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:12Z","lastTransitionTime":"2025-12-01T17:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.710344 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.710412 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.710424 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.710450 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.710465 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:12Z","lastTransitionTime":"2025-12-01T17:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.813492 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.813548 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.813560 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.813579 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.813590 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:12Z","lastTransitionTime":"2025-12-01T17:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.916639 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.916688 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.916696 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.916712 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:12 crc kubenswrapper[4868]: I1201 17:26:12.916721 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:12Z","lastTransitionTime":"2025-12-01T17:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.018850 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.018910 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.018923 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.018957 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.018970 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:13Z","lastTransitionTime":"2025-12-01T17:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.122729 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.122816 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.122829 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.122850 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.122862 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:13Z","lastTransitionTime":"2025-12-01T17:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.172002 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.172125 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:13 crc kubenswrapper[4868]: E1201 17:26:13.172205 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:13 crc kubenswrapper[4868]: E1201 17:26:13.172312 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.226162 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.226234 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.226261 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.226294 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.226319 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:13Z","lastTransitionTime":"2025-12-01T17:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.329381 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.329434 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.329450 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.329471 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.329483 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:13Z","lastTransitionTime":"2025-12-01T17:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.432001 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.432096 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.432116 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.432145 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.432169 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:13Z","lastTransitionTime":"2025-12-01T17:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.534749 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.534791 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.534808 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.534830 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.534843 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:13Z","lastTransitionTime":"2025-12-01T17:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.637488 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.637530 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.637539 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.637554 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.637564 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:13Z","lastTransitionTime":"2025-12-01T17:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.739894 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.739963 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.739977 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.740000 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.740016 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:13Z","lastTransitionTime":"2025-12-01T17:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.842307 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.842390 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.842413 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.842447 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.842471 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:13Z","lastTransitionTime":"2025-12-01T17:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.945726 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.945812 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.945831 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.945866 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:13 crc kubenswrapper[4868]: I1201 17:26:13.945885 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:13Z","lastTransitionTime":"2025-12-01T17:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.049195 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.049273 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.049293 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.049323 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.049344 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:14Z","lastTransitionTime":"2025-12-01T17:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.152704 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.152792 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.152811 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.152837 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.152855 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:14Z","lastTransitionTime":"2025-12-01T17:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.171498 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:14 crc kubenswrapper[4868]: E1201 17:26:14.171723 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.172125 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:14 crc kubenswrapper[4868]: E1201 17:26:14.172383 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.256171 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.256236 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.256247 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.256272 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.256285 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:14Z","lastTransitionTime":"2025-12-01T17:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.359755 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.359807 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.359819 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.359839 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.359852 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:14Z","lastTransitionTime":"2025-12-01T17:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.463287 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.463391 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.463419 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.463457 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.463480 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:14Z","lastTransitionTime":"2025-12-01T17:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.566544 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.566603 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.566620 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.566637 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.566648 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:14Z","lastTransitionTime":"2025-12-01T17:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.669901 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.669964 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.669983 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.670003 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.670016 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:14Z","lastTransitionTime":"2025-12-01T17:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.772809 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.772855 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.772863 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.772879 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.772890 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:14Z","lastTransitionTime":"2025-12-01T17:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.875084 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.875142 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.875156 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.875177 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.875190 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:14Z","lastTransitionTime":"2025-12-01T17:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.977920 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.977982 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.977991 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.978008 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:14 crc kubenswrapper[4868]: I1201 17:26:14.978019 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:14Z","lastTransitionTime":"2025-12-01T17:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.080874 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.080967 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.080985 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.081012 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.081046 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:15Z","lastTransitionTime":"2025-12-01T17:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.171371 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.171485 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:15 crc kubenswrapper[4868]: E1201 17:26:15.171624 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:15 crc kubenswrapper[4868]: E1201 17:26:15.171782 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.183270 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.183328 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.183348 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.183375 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.183397 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:15Z","lastTransitionTime":"2025-12-01T17:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.286520 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.286564 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.286573 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.286588 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.286597 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:15Z","lastTransitionTime":"2025-12-01T17:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.391396 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.391443 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.391455 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.391474 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.391489 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:15Z","lastTransitionTime":"2025-12-01T17:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.494218 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.494272 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.494324 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.494347 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.494360 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:15Z","lastTransitionTime":"2025-12-01T17:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.597300 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.597383 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.597411 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.597438 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.597457 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:15Z","lastTransitionTime":"2025-12-01T17:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.700487 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.700540 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.700555 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.700575 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.700590 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:15Z","lastTransitionTime":"2025-12-01T17:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.803412 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.803478 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.803497 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.803527 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.803546 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:15Z","lastTransitionTime":"2025-12-01T17:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.906593 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.906674 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.906698 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.906730 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:15 crc kubenswrapper[4868]: I1201 17:26:15.906754 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:15Z","lastTransitionTime":"2025-12-01T17:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.010439 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.010491 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.010501 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.010520 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.010560 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:16Z","lastTransitionTime":"2025-12-01T17:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.113271 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.113327 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.113339 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.113357 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.113367 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:16Z","lastTransitionTime":"2025-12-01T17:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.171278 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.171444 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:16 crc kubenswrapper[4868]: E1201 17:26:16.171517 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:16 crc kubenswrapper[4868]: E1201 17:26:16.171693 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.215403 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.215455 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.215468 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.215488 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.215506 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:16Z","lastTransitionTime":"2025-12-01T17:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.319206 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.319257 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.319266 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.319282 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.319292 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:16Z","lastTransitionTime":"2025-12-01T17:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.422313 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.422368 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.422381 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.422402 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.422413 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:16Z","lastTransitionTime":"2025-12-01T17:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.493655 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs\") pod \"network-metrics-daemon-nxfkx\" (UID: \"4afd9a0c-e8c7-42ec-9771-036c3441cbce\") " pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:16 crc kubenswrapper[4868]: E1201 17:26:16.493829 4868 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:26:16 crc kubenswrapper[4868]: E1201 17:26:16.493898 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs podName:4afd9a0c-e8c7-42ec-9771-036c3441cbce nodeName:}" failed. No retries permitted until 2025-12-01 17:26:24.493878458 +0000 UTC m=+56.864988879 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs") pod "network-metrics-daemon-nxfkx" (UID: "4afd9a0c-e8c7-42ec-9771-036c3441cbce") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.524702 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.524766 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.524776 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.524800 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.524810 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:16Z","lastTransitionTime":"2025-12-01T17:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.627418 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.627457 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.627470 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.627492 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.627509 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:16Z","lastTransitionTime":"2025-12-01T17:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.730427 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.730481 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.730492 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.730512 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.730523 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:16Z","lastTransitionTime":"2025-12-01T17:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.834046 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.834101 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.834119 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.834145 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.834162 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:16Z","lastTransitionTime":"2025-12-01T17:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.937457 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.937514 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.937533 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.937556 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:16 crc kubenswrapper[4868]: I1201 17:26:16.937575 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:16Z","lastTransitionTime":"2025-12-01T17:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.041162 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.041224 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.041242 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.041266 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.041283 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:17Z","lastTransitionTime":"2025-12-01T17:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.145069 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.145113 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.145125 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.145143 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.145156 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:17Z","lastTransitionTime":"2025-12-01T17:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.171517 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:17 crc kubenswrapper[4868]: E1201 17:26:17.171697 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.171909 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:17 crc kubenswrapper[4868]: E1201 17:26:17.172000 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.247629 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.247660 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.247668 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.247682 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.247691 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:17Z","lastTransitionTime":"2025-12-01T17:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.350223 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.350276 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.350288 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.350310 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.350321 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:17Z","lastTransitionTime":"2025-12-01T17:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.453156 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.453217 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.453232 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.453258 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.453281 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:17Z","lastTransitionTime":"2025-12-01T17:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.556574 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.556694 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.556711 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.556735 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.556748 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:17Z","lastTransitionTime":"2025-12-01T17:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.659177 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.659248 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.659258 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.659276 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.659289 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:17Z","lastTransitionTime":"2025-12-01T17:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.762083 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.762118 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.762128 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.762143 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.762155 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:17Z","lastTransitionTime":"2025-12-01T17:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.864312 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.864373 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.864391 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.864415 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.864465 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:17Z","lastTransitionTime":"2025-12-01T17:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.968197 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.968252 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.968273 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.968297 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:17 crc kubenswrapper[4868]: I1201 17:26:17.968316 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:17Z","lastTransitionTime":"2025-12-01T17:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.071907 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.071987 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.072006 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.072026 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.072040 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:18Z","lastTransitionTime":"2025-12-01T17:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.171634 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.171816 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:18 crc kubenswrapper[4868]: E1201 17:26:18.171852 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:18 crc kubenswrapper[4868]: E1201 17:26:18.172107 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.175145 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.175183 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.175193 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.175207 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.175220 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:18Z","lastTransitionTime":"2025-12-01T17:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.187125 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.200121 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.222008 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://213c840254d569cf46c44f2ebb230e3f03490cd511b1cf5630cf70af816e6f32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"tory.go:141\\\\nI1201 17:26:07.251211 6159 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:07.251238 6159 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:07.251282 6159 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.251301 6159 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.251394 6159 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:07.253796 6159 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 17:26:07.253814 6159 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 17:26:07.253846 6159 factory.go:656] Stopping watch factory\\\\nI1201 17:26:07.253852 6159 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 17:26:07.253870 6159 ovnkube.go:599] Stopped ovnkube\\\\nI1201 17\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:09Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.556807 6333 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.556910 6333 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.557032 6333 factory.go:656] Stopping watch factory\\\\nI1201 17:26:09.557052 6333 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557341 6333 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557758 6333 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:09.559974 6333 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1201 17:26:09.559996 6333 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1201 17:26:09.560056 6333 ovnkube.go:599] Stopped ovnkube\\\\nI1201 17:26:09.560088 6333 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1201 17:26:09.560163 6333 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.234527 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.246478 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.261844 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.277208 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.277242 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.277251 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.277267 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.277277 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:18Z","lastTransitionTime":"2025-12-01T17:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.283526 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.293606 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.306169 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.319486 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.332871 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.345168 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.356595 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.368336 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.379340 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.379373 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.379407 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.379425 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.379437 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:18Z","lastTransitionTime":"2025-12-01T17:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.381309 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.393338 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:18Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.482093 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.482529 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.482614 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.482709 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.482788 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:18Z","lastTransitionTime":"2025-12-01T17:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.585729 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.585771 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.585782 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.585796 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.585807 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:18Z","lastTransitionTime":"2025-12-01T17:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.688661 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.688984 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.689090 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.689190 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.689272 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:18Z","lastTransitionTime":"2025-12-01T17:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.791532 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.791576 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.791593 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.791621 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.791633 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:18Z","lastTransitionTime":"2025-12-01T17:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.893806 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.893849 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.893858 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.893875 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.893886 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:18Z","lastTransitionTime":"2025-12-01T17:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.996817 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.996855 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.996864 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.996879 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:18 crc kubenswrapper[4868]: I1201 17:26:18.996890 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:18Z","lastTransitionTime":"2025-12-01T17:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.022379 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.022539 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:19 crc kubenswrapper[4868]: E1201 17:26:19.022649 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:26:51.022615199 +0000 UTC m=+83.393725610 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:26:19 crc kubenswrapper[4868]: E1201 17:26:19.022667 4868 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:26:19 crc kubenswrapper[4868]: E1201 17:26:19.022730 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:26:51.022713812 +0000 UTC m=+83.393824223 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.022749 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:19 crc kubenswrapper[4868]: E1201 17:26:19.022818 4868 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:26:19 crc kubenswrapper[4868]: E1201 17:26:19.022840 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:26:51.022834385 +0000 UTC m=+83.393944786 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.099914 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.099974 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.099985 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.100005 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.100015 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:19Z","lastTransitionTime":"2025-12-01T17:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.123687 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.123742 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:19 crc kubenswrapper[4868]: E1201 17:26:19.123877 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:26:19 crc kubenswrapper[4868]: E1201 17:26:19.123921 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:26:19 crc kubenswrapper[4868]: E1201 17:26:19.123934 4868 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:26:19 crc kubenswrapper[4868]: E1201 17:26:19.123888 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:26:19 crc kubenswrapper[4868]: E1201 17:26:19.123982 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:26:19 crc kubenswrapper[4868]: E1201 17:26:19.123994 4868 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:26:19 crc kubenswrapper[4868]: E1201 17:26:19.124017 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 17:26:51.123997317 +0000 UTC m=+83.495107728 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:26:19 crc kubenswrapper[4868]: E1201 17:26:19.124045 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 17:26:51.124028027 +0000 UTC m=+83.495138438 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.171281 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.171286 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:19 crc kubenswrapper[4868]: E1201 17:26:19.171414 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:19 crc kubenswrapper[4868]: E1201 17:26:19.171544 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.202093 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.202123 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.202134 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.202147 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.202158 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:19Z","lastTransitionTime":"2025-12-01T17:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.304729 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.304782 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.304796 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.304821 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.304832 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:19Z","lastTransitionTime":"2025-12-01T17:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.407613 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.407674 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.407687 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.407709 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.407733 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:19Z","lastTransitionTime":"2025-12-01T17:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.511392 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.511463 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.511489 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.511523 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.511551 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:19Z","lastTransitionTime":"2025-12-01T17:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.615008 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.615055 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.615070 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.615097 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.615112 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:19Z","lastTransitionTime":"2025-12-01T17:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.717426 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.717475 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.717491 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.717516 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.717532 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:19Z","lastTransitionTime":"2025-12-01T17:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.821151 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.821685 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.821810 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.821975 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.822184 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:19Z","lastTransitionTime":"2025-12-01T17:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.925352 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.925817 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.926185 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.926502 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:19 crc kubenswrapper[4868]: I1201 17:26:19.926878 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:19Z","lastTransitionTime":"2025-12-01T17:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.030836 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.030904 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.030915 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.030934 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.030961 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:20Z","lastTransitionTime":"2025-12-01T17:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.134104 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.134162 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.134177 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.134197 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.134207 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:20Z","lastTransitionTime":"2025-12-01T17:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.171656 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.171721 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:20 crc kubenswrapper[4868]: E1201 17:26:20.171816 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:20 crc kubenswrapper[4868]: E1201 17:26:20.171956 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.236753 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.236795 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.236804 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.236821 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.236833 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:20Z","lastTransitionTime":"2025-12-01T17:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.338687 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.338731 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.338739 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.338756 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.338768 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:20Z","lastTransitionTime":"2025-12-01T17:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.440390 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.440441 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.440459 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.440482 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.440499 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:20Z","lastTransitionTime":"2025-12-01T17:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.542705 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.542748 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.542780 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.542801 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.542812 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:20Z","lastTransitionTime":"2025-12-01T17:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.644883 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.644933 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.644968 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.645002 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.645016 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:20Z","lastTransitionTime":"2025-12-01T17:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.747564 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.747598 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.747609 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.747624 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.747635 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:20Z","lastTransitionTime":"2025-12-01T17:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.850085 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.850141 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.850154 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.850174 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.850185 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:20Z","lastTransitionTime":"2025-12-01T17:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.952492 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.952814 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.952885 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.952976 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:20 crc kubenswrapper[4868]: I1201 17:26:20.953046 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:20Z","lastTransitionTime":"2025-12-01T17:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.055188 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.055224 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.055232 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.055250 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.055259 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:21Z","lastTransitionTime":"2025-12-01T17:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.157769 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.157818 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.157832 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.157868 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.157880 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:21Z","lastTransitionTime":"2025-12-01T17:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.171288 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.171325 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:21 crc kubenswrapper[4868]: E1201 17:26:21.171425 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:21 crc kubenswrapper[4868]: E1201 17:26:21.171542 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.259878 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.259922 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.259932 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.259969 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.259980 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:21Z","lastTransitionTime":"2025-12-01T17:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.361884 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.361965 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.361982 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.362000 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.362012 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:21Z","lastTransitionTime":"2025-12-01T17:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.464689 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.465047 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.465155 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.465266 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.465369 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:21Z","lastTransitionTime":"2025-12-01T17:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.568235 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.568527 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.568625 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.568710 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.568798 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:21Z","lastTransitionTime":"2025-12-01T17:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.671381 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.671428 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.671439 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.671457 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.671469 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:21Z","lastTransitionTime":"2025-12-01T17:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.773693 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.773736 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.773749 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.773765 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.773776 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:21Z","lastTransitionTime":"2025-12-01T17:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.875795 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.875830 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.875838 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.875853 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.875862 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:21Z","lastTransitionTime":"2025-12-01T17:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.978817 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.978867 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.978881 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.978921 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:21 crc kubenswrapper[4868]: I1201 17:26:21.978955 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:21Z","lastTransitionTime":"2025-12-01T17:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.081313 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.081360 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.081372 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.081394 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.081405 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:22Z","lastTransitionTime":"2025-12-01T17:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.171442 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.171570 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:22 crc kubenswrapper[4868]: E1201 17:26:22.171588 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:22 crc kubenswrapper[4868]: E1201 17:26:22.171718 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.183644 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.183687 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.183696 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.183714 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.183724 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:22Z","lastTransitionTime":"2025-12-01T17:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.285678 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.285726 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.285737 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.285755 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.285768 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:22Z","lastTransitionTime":"2025-12-01T17:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.388673 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.388928 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.389062 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.389173 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.389234 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:22Z","lastTransitionTime":"2025-12-01T17:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.392751 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.393575 4868 scope.go:117] "RemoveContainer" containerID="48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.419540 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.435368 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.448403 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.455394 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.455439 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.455451 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.455469 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.455478 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:22Z","lastTransitionTime":"2025-12-01T17:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.461725 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: E1201 17:26:22.469122 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.473334 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.473365 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.473374 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.473391 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.473443 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:22Z","lastTransitionTime":"2025-12-01T17:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.475626 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: E1201 17:26:22.485581 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.487857 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.489964 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.489990 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.490001 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.490019 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.490034 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:22Z","lastTransitionTime":"2025-12-01T17:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.499469 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: E1201 17:26:22.508267 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.511352 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.511386 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.511396 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.511412 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.511423 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:22Z","lastTransitionTime":"2025-12-01T17:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.519786 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: E1201 17:26:22.530730 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.535598 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.535640 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.535649 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.535665 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.535677 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:22Z","lastTransitionTime":"2025-12-01T17:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.540904 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:09Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.556807 6333 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.556910 6333 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.557032 6333 factory.go:656] Stopping watch factory\\\\nI1201 17:26:09.557052 6333 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557341 6333 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557758 6333 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:09.559974 6333 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1201 17:26:09.559996 6333 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1201 17:26:09.560056 6333 ovnkube.go:599] Stopped ovnkube\\\\nI1201 17:26:09.560088 6333 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1201 17:26:09.560163 6333 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-q9z6l_openshift-ovn-kubernetes(5baca705-90cc-4198-b641-214a2b915086)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: E1201 17:26:22.548150 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: E1201 17:26:22.548291 4868 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.550160 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.550192 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.550202 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.550222 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.550234 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:22Z","lastTransitionTime":"2025-12-01T17:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.554445 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.567531 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.577555 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.591620 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.605876 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.620903 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.633555 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:22Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.652179 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.652228 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.652242 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.652294 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.652309 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:22Z","lastTransitionTime":"2025-12-01T17:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.754365 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.754416 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.754427 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.754443 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.754453 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:22Z","lastTransitionTime":"2025-12-01T17:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.858545 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.858580 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.858592 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.858612 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.858627 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:22Z","lastTransitionTime":"2025-12-01T17:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.962247 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.962282 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.962291 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.962305 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:22 crc kubenswrapper[4868]: I1201 17:26:22.962314 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:22Z","lastTransitionTime":"2025-12-01T17:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.007590 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.023617 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.026354 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.041724 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.051769 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.064204 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.065051 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.065105 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.065117 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.065137 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.065150 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:23Z","lastTransitionTime":"2025-12-01T17:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.082377 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:09Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.556807 6333 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.556910 6333 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.557032 6333 factory.go:656] Stopping watch factory\\\\nI1201 17:26:09.557052 6333 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557341 6333 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557758 6333 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:09.559974 6333 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1201 17:26:09.559996 6333 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1201 17:26:09.560056 6333 ovnkube.go:599] Stopped ovnkube\\\\nI1201 17:26:09.560088 6333 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1201 17:26:09.560163 6333 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-q9z6l_openshift-ovn-kubernetes(5baca705-90cc-4198-b641-214a2b915086)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.093275 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.104270 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.117350 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.129304 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.146393 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.159800 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.170897 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.170907 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:23 crc kubenswrapper[4868]: E1201 17:26:23.171071 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:23 crc kubenswrapper[4868]: E1201 17:26:23.171155 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.171967 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.172015 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.172026 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.172044 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.172056 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:23Z","lastTransitionTime":"2025-12-01T17:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.173229 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.187400 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.199913 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.210983 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.226347 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.273927 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.273987 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.274000 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.274019 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.274031 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:23Z","lastTransitionTime":"2025-12-01T17:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.375923 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.376027 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.376037 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.376054 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.376068 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:23Z","lastTransitionTime":"2025-12-01T17:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.486338 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.486396 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.486410 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.486435 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.486458 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:23Z","lastTransitionTime":"2025-12-01T17:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.591633 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.591663 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.591674 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.591693 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.591706 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:23Z","lastTransitionTime":"2025-12-01T17:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.592372 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovnkube-controller/1.log" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.595211 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerStarted","Data":"d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb"} Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.608161 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cc7d784-77de-4109-ba8c-192bb74290f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c84b10381473d660a8159f66c6c2f012e0d46e588ef7f19a0f6950edc94806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://120b3cade14d3c00f0a22c5f4db0c5500a669bb07a267b65bf7a386938467f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://865f0cda78b8e8ef0cd3d52fd8c4baa2e0c0f67e9d2d89752ac7a7f5e5df0a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.622406 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.637145 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.652726 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.670771 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.692997 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.694408 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.694456 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.694465 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.694482 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.694494 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:23Z","lastTransitionTime":"2025-12-01T17:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.708879 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.722863 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.759366 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.780562 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.797257 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.797298 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.797307 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.797324 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.797334 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:23Z","lastTransitionTime":"2025-12-01T17:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.806900 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.819746 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.829769 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.844407 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.863304 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:09Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.556807 6333 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.556910 6333 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.557032 6333 factory.go:656] Stopping watch factory\\\\nI1201 17:26:09.557052 6333 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557341 6333 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557758 6333 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:09.559974 6333 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1201 17:26:09.559996 6333 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1201 17:26:09.560056 6333 ovnkube.go:599] Stopped ovnkube\\\\nI1201 17:26:09.560088 6333 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1201 17:26:09.560163 6333 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.877171 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.890253 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:23Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.900178 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.900247 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.900279 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.900304 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:23 crc kubenswrapper[4868]: I1201 17:26:23.900319 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:23Z","lastTransitionTime":"2025-12-01T17:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.002367 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.002410 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.002423 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.002440 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.002454 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:24Z","lastTransitionTime":"2025-12-01T17:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.104510 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.104816 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.104829 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.104846 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.104856 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:24Z","lastTransitionTime":"2025-12-01T17:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.171531 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.171638 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:24 crc kubenswrapper[4868]: E1201 17:26:24.172133 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:24 crc kubenswrapper[4868]: E1201 17:26:24.172271 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.207866 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.207903 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.207916 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.207933 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.207965 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:24Z","lastTransitionTime":"2025-12-01T17:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.310587 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.310647 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.310665 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.311049 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.311085 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:24Z","lastTransitionTime":"2025-12-01T17:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.413822 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.413870 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.413883 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.413900 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.413911 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:24Z","lastTransitionTime":"2025-12-01T17:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.516690 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.516996 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.517091 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.517161 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.517247 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:24Z","lastTransitionTime":"2025-12-01T17:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.581350 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs\") pod \"network-metrics-daemon-nxfkx\" (UID: \"4afd9a0c-e8c7-42ec-9771-036c3441cbce\") " pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:24 crc kubenswrapper[4868]: E1201 17:26:24.581579 4868 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:26:24 crc kubenswrapper[4868]: E1201 17:26:24.581654 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs podName:4afd9a0c-e8c7-42ec-9771-036c3441cbce nodeName:}" failed. No retries permitted until 2025-12-01 17:26:40.581631528 +0000 UTC m=+72.952741949 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs") pod "network-metrics-daemon-nxfkx" (UID: "4afd9a0c-e8c7-42ec-9771-036c3441cbce") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.600633 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovnkube-controller/2.log" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.601437 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovnkube-controller/1.log" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.604993 4868 generic.go:334] "Generic (PLEG): container finished" podID="5baca705-90cc-4198-b641-214a2b915086" containerID="d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb" exitCode=1 Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.605082 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerDied","Data":"d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb"} Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.605173 4868 scope.go:117] "RemoveContainer" containerID="48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.606150 4868 scope.go:117] "RemoveContainer" containerID="d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb" Dec 01 17:26:24 crc kubenswrapper[4868]: E1201 17:26:24.606402 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-q9z6l_openshift-ovn-kubernetes(5baca705-90cc-4198-b641-214a2b915086)\"" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podUID="5baca705-90cc-4198-b641-214a2b915086" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.621351 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.621390 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.621398 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.621414 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.621427 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:24Z","lastTransitionTime":"2025-12-01T17:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.622911 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.641448 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.655758 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.680827 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.698040 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.721814 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:09Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.556807 6333 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.556910 6333 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.557032 6333 factory.go:656] Stopping watch factory\\\\nI1201 17:26:09.557052 6333 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557341 6333 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557758 6333 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:09.559974 6333 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1201 17:26:09.559996 6333 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1201 17:26:09.560056 6333 ovnkube.go:599] Stopped ovnkube\\\\nI1201 17:26:09.560088 6333 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1201 17:26:09.560163 6333 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:24Z\\\",\\\"message\\\":\\\"io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00784edfb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:mhc-mtrc,Protocol:TCP,Port:8444,TargetPort:{1 0 mhc-mtrc},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: controller,},ClusterIP:10.217.4.167,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.167],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1201 17:26:24.043188 6530 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.724306 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.724365 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.724379 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.724399 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.724414 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:24Z","lastTransitionTime":"2025-12-01T17:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.736749 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.749806 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.765007 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.780196 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.792080 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.810304 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.825349 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.826435 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.826479 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.826489 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.826508 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.826517 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:24Z","lastTransitionTime":"2025-12-01T17:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.843169 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cc7d784-77de-4109-ba8c-192bb74290f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c84b10381473d660a8159f66c6c2f012e0d46e588ef7f19a0f6950edc94806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://120b3cade14d3c00f0a22c5f4db0c5500a669bb07a267b65bf7a386938467f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://865f0cda78b8e8ef0cd3d52fd8c4baa2e0c0f67e9d2d89752ac7a7f5e5df0a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.860013 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.878609 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.895362 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:24Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.929844 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.929912 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.929925 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.929982 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:24 crc kubenswrapper[4868]: I1201 17:26:24.930003 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:24Z","lastTransitionTime":"2025-12-01T17:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.032591 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.032639 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.032650 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.032667 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.032678 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:25Z","lastTransitionTime":"2025-12-01T17:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.135212 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.135261 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.135274 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.135295 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.135308 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:25Z","lastTransitionTime":"2025-12-01T17:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.171851 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.171916 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:25 crc kubenswrapper[4868]: E1201 17:26:25.172069 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:25 crc kubenswrapper[4868]: E1201 17:26:25.172201 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.238432 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.238527 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.238538 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.238555 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.238565 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:25Z","lastTransitionTime":"2025-12-01T17:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.341671 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.341722 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.341731 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.341758 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.341769 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:25Z","lastTransitionTime":"2025-12-01T17:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.443952 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.443999 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.444011 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.444028 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.444038 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:25Z","lastTransitionTime":"2025-12-01T17:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.546888 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.546932 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.546954 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.546977 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.546988 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:25Z","lastTransitionTime":"2025-12-01T17:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.610617 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovnkube-controller/2.log" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.652545 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.652591 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.652602 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.652631 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.652645 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:25Z","lastTransitionTime":"2025-12-01T17:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.755688 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.755761 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.755775 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.755793 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.755804 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:25Z","lastTransitionTime":"2025-12-01T17:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.858683 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.858733 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.858745 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.858767 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.858778 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:25Z","lastTransitionTime":"2025-12-01T17:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.961987 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.962032 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.962043 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.962060 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:25 crc kubenswrapper[4868]: I1201 17:26:25.962071 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:25Z","lastTransitionTime":"2025-12-01T17:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.064641 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.064671 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.064679 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.064694 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.064704 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:26Z","lastTransitionTime":"2025-12-01T17:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.166771 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.166823 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.166835 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.166854 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.166867 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:26Z","lastTransitionTime":"2025-12-01T17:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.171158 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.171206 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:26 crc kubenswrapper[4868]: E1201 17:26:26.171287 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:26 crc kubenswrapper[4868]: E1201 17:26:26.171380 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.270043 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.270090 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.270101 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.270116 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.270126 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:26Z","lastTransitionTime":"2025-12-01T17:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.372569 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.372614 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.372624 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.372642 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.372652 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:26Z","lastTransitionTime":"2025-12-01T17:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.475294 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.475334 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.475345 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.475363 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.475375 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:26Z","lastTransitionTime":"2025-12-01T17:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.577775 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.577808 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.577817 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.577832 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.577843 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:26Z","lastTransitionTime":"2025-12-01T17:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.680048 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.680084 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.680093 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.680109 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.680118 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:26Z","lastTransitionTime":"2025-12-01T17:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.782043 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.782093 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.782108 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.782125 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.782137 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:26Z","lastTransitionTime":"2025-12-01T17:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.884842 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.884882 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.884894 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.884912 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.884923 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:26Z","lastTransitionTime":"2025-12-01T17:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.988444 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.988660 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.988814 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.988993 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:26 crc kubenswrapper[4868]: I1201 17:26:26.989387 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:26Z","lastTransitionTime":"2025-12-01T17:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.091532 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.091579 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.091593 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.091610 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.091622 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:27Z","lastTransitionTime":"2025-12-01T17:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.171889 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:27 crc kubenswrapper[4868]: E1201 17:26:27.172058 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.171894 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:27 crc kubenswrapper[4868]: E1201 17:26:27.172310 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.193751 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.193799 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.193812 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.193833 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.193845 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:27Z","lastTransitionTime":"2025-12-01T17:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.296850 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.296916 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.296931 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.296978 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.296993 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:27Z","lastTransitionTime":"2025-12-01T17:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.399849 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.399895 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.399909 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.399928 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.399955 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:27Z","lastTransitionTime":"2025-12-01T17:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.502796 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.502851 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.502864 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.502887 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.502903 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:27Z","lastTransitionTime":"2025-12-01T17:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.606094 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.606143 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.606165 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.606183 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.606195 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:27Z","lastTransitionTime":"2025-12-01T17:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.709051 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.709088 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.709102 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.709119 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.709129 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:27Z","lastTransitionTime":"2025-12-01T17:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.811791 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.811841 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.811857 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.811876 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.811887 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:27Z","lastTransitionTime":"2025-12-01T17:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.914209 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.914260 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.914273 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.914296 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:27 crc kubenswrapper[4868]: I1201 17:26:27.914311 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:27Z","lastTransitionTime":"2025-12-01T17:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.016876 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.017358 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.017485 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.017623 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.017924 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:28Z","lastTransitionTime":"2025-12-01T17:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.120763 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.120824 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.120838 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.120858 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.120876 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:28Z","lastTransitionTime":"2025-12-01T17:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.171727 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.171984 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:28 crc kubenswrapper[4868]: E1201 17:26:28.172641 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:28 crc kubenswrapper[4868]: E1201 17:26:28.173134 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.190497 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.204514 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.225670 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.225751 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.225767 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.225793 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.225808 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:28Z","lastTransitionTime":"2025-12-01T17:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.228976 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.251932 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:09Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.556807 6333 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.556910 6333 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.557032 6333 factory.go:656] Stopping watch factory\\\\nI1201 17:26:09.557052 6333 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557341 6333 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557758 6333 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:09.559974 6333 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1201 17:26:09.559996 6333 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1201 17:26:09.560056 6333 ovnkube.go:599] Stopped ovnkube\\\\nI1201 17:26:09.560088 6333 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1201 17:26:09.560163 6333 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:24Z\\\",\\\"message\\\":\\\"io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00784edfb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:mhc-mtrc,Protocol:TCP,Port:8444,TargetPort:{1 0 mhc-mtrc},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: controller,},ClusterIP:10.217.4.167,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.167],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1201 17:26:24.043188 6530 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.267863 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.284389 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.302033 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.318169 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.329108 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.329151 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.329168 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.329191 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.329208 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:28Z","lastTransitionTime":"2025-12-01T17:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.334245 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.348108 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.364651 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.377559 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cc7d784-77de-4109-ba8c-192bb74290f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c84b10381473d660a8159f66c6c2f012e0d46e588ef7f19a0f6950edc94806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://120b3cade14d3c00f0a22c5f4db0c5500a669bb07a267b65bf7a386938467f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://865f0cda78b8e8ef0cd3d52fd8c4baa2e0c0f67e9d2d89752ac7a7f5e5df0a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.395556 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.411763 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.425274 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.431814 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.431860 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.431875 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.431900 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.431917 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:28Z","lastTransitionTime":"2025-12-01T17:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.440555 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.456708 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:28Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.534409 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.534810 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.534826 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.534852 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.534868 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:28Z","lastTransitionTime":"2025-12-01T17:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.637659 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.637718 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.637731 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.637750 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.637764 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:28Z","lastTransitionTime":"2025-12-01T17:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.740750 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.740824 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.740836 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.740855 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.740868 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:28Z","lastTransitionTime":"2025-12-01T17:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.843675 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.843781 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.843808 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.843838 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.843860 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:28Z","lastTransitionTime":"2025-12-01T17:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.945984 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.946050 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.946060 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.946080 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:28 crc kubenswrapper[4868]: I1201 17:26:28.946092 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:28Z","lastTransitionTime":"2025-12-01T17:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.048748 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.048805 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.048820 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.048843 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.048858 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:29Z","lastTransitionTime":"2025-12-01T17:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.157031 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.157089 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.157110 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.157129 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.157143 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:29Z","lastTransitionTime":"2025-12-01T17:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.170871 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:29 crc kubenswrapper[4868]: E1201 17:26:29.171056 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.170871 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:29 crc kubenswrapper[4868]: E1201 17:26:29.171146 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.260208 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.260255 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.260267 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.260288 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.260301 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:29Z","lastTransitionTime":"2025-12-01T17:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.363230 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.363282 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.363292 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.363311 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.363323 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:29Z","lastTransitionTime":"2025-12-01T17:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.465639 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.465682 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.465690 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.465704 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.465713 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:29Z","lastTransitionTime":"2025-12-01T17:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.568538 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.568961 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.569097 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.569202 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.569372 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:29Z","lastTransitionTime":"2025-12-01T17:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.672227 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.672281 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.672295 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.672315 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.672332 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:29Z","lastTransitionTime":"2025-12-01T17:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.775625 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.775676 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.775686 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.775703 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.775714 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:29Z","lastTransitionTime":"2025-12-01T17:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.878193 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.878256 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.878271 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.878290 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.878304 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:29Z","lastTransitionTime":"2025-12-01T17:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.981827 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.981896 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.981912 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.981966 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:29 crc kubenswrapper[4868]: I1201 17:26:29.981987 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:29Z","lastTransitionTime":"2025-12-01T17:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.083887 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.083924 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.083965 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.083983 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.083995 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:30Z","lastTransitionTime":"2025-12-01T17:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.171495 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.171604 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:30 crc kubenswrapper[4868]: E1201 17:26:30.171667 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:30 crc kubenswrapper[4868]: E1201 17:26:30.171807 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.186389 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.186432 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.186446 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.186465 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.186478 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:30Z","lastTransitionTime":"2025-12-01T17:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.289502 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.289582 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.289602 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.289633 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.289653 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:30Z","lastTransitionTime":"2025-12-01T17:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.392597 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.392625 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.392651 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.392668 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.392677 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:30Z","lastTransitionTime":"2025-12-01T17:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.495696 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.495728 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.495738 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.495753 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.495763 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:30Z","lastTransitionTime":"2025-12-01T17:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.598474 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.598543 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.598553 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.598574 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.598609 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:30Z","lastTransitionTime":"2025-12-01T17:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.702593 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.702645 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.702660 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.702683 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.702700 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:30Z","lastTransitionTime":"2025-12-01T17:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.805671 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.806395 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.806423 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.806455 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.806477 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:30Z","lastTransitionTime":"2025-12-01T17:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.908781 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.908850 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.908868 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.908888 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:30 crc kubenswrapper[4868]: I1201 17:26:30.909121 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:30Z","lastTransitionTime":"2025-12-01T17:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.012747 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.012813 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.012826 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.012847 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.012859 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:31Z","lastTransitionTime":"2025-12-01T17:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.116141 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.116197 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.116212 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.116234 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.116245 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:31Z","lastTransitionTime":"2025-12-01T17:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.171739 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:31 crc kubenswrapper[4868]: E1201 17:26:31.171961 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.171750 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:31 crc kubenswrapper[4868]: E1201 17:26:31.172224 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.219545 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.219591 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.219601 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.219621 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.219633 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:31Z","lastTransitionTime":"2025-12-01T17:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.322819 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.322876 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.322885 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.322902 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.322914 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:31Z","lastTransitionTime":"2025-12-01T17:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.425371 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.425419 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.425430 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.425449 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.425464 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:31Z","lastTransitionTime":"2025-12-01T17:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.529393 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.529437 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.529449 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.529466 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.529478 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:31Z","lastTransitionTime":"2025-12-01T17:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.632502 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.632555 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.632564 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.632589 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.632600 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:31Z","lastTransitionTime":"2025-12-01T17:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.735332 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.735380 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.735389 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.735404 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.735415 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:31Z","lastTransitionTime":"2025-12-01T17:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.838170 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.838246 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.838260 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.838282 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.838298 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:31Z","lastTransitionTime":"2025-12-01T17:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.941349 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.941393 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.941406 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.941430 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:31 crc kubenswrapper[4868]: I1201 17:26:31.941443 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:31Z","lastTransitionTime":"2025-12-01T17:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.043770 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.043816 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.043827 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.043845 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.043855 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:32Z","lastTransitionTime":"2025-12-01T17:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.147012 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.147071 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.147096 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.147119 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.147132 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:32Z","lastTransitionTime":"2025-12-01T17:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.171311 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:32 crc kubenswrapper[4868]: E1201 17:26:32.171444 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.171707 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:32 crc kubenswrapper[4868]: E1201 17:26:32.171782 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.249634 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.249719 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.249730 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.249746 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.249758 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:32Z","lastTransitionTime":"2025-12-01T17:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.352542 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.352601 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.352612 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.352633 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.352648 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:32Z","lastTransitionTime":"2025-12-01T17:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.454997 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.455046 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.455056 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.455077 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.455090 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:32Z","lastTransitionTime":"2025-12-01T17:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.557394 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.557443 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.557457 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.557477 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.557489 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:32Z","lastTransitionTime":"2025-12-01T17:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.659980 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.660019 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.660029 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.660043 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.660054 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:32Z","lastTransitionTime":"2025-12-01T17:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.707356 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.707421 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.707436 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.707460 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.707475 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:32Z","lastTransitionTime":"2025-12-01T17:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:32 crc kubenswrapper[4868]: E1201 17:26:32.721311 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:32Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.725193 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.725250 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.725265 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.725287 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.725301 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:32Z","lastTransitionTime":"2025-12-01T17:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:32 crc kubenswrapper[4868]: E1201 17:26:32.740649 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:32Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.745466 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.745526 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.745541 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.745565 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.745582 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:32Z","lastTransitionTime":"2025-12-01T17:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:32 crc kubenswrapper[4868]: E1201 17:26:32.762166 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:32Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.766548 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.766609 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.766626 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.766651 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.766670 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:32Z","lastTransitionTime":"2025-12-01T17:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:32 crc kubenswrapper[4868]: E1201 17:26:32.786903 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:32Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.792442 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.792480 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.792495 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.792515 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.792529 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:32Z","lastTransitionTime":"2025-12-01T17:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:32 crc kubenswrapper[4868]: E1201 17:26:32.809132 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:32Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:32 crc kubenswrapper[4868]: E1201 17:26:32.809262 4868 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.811761 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.811801 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.811816 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.811839 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.811854 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:32Z","lastTransitionTime":"2025-12-01T17:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.914204 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.914246 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.914258 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.914277 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:32 crc kubenswrapper[4868]: I1201 17:26:32.914286 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:32Z","lastTransitionTime":"2025-12-01T17:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.017706 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.017767 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.017783 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.017801 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.017814 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:33Z","lastTransitionTime":"2025-12-01T17:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.121134 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.121182 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.121195 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.121216 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.121233 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:33Z","lastTransitionTime":"2025-12-01T17:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.171419 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.171454 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:33 crc kubenswrapper[4868]: E1201 17:26:33.171610 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:33 crc kubenswrapper[4868]: E1201 17:26:33.171718 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.223906 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.224002 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.224022 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.224053 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.224075 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:33Z","lastTransitionTime":"2025-12-01T17:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.328247 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.328425 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.328495 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.328641 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.328734 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:33Z","lastTransitionTime":"2025-12-01T17:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.431534 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.431592 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.431604 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.431624 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.431640 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:33Z","lastTransitionTime":"2025-12-01T17:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.534365 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.534409 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.534422 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.534444 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.534459 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:33Z","lastTransitionTime":"2025-12-01T17:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.637361 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.637402 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.637412 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.637426 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.637436 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:33Z","lastTransitionTime":"2025-12-01T17:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.740412 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.740469 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.740481 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.740500 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.740512 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:33Z","lastTransitionTime":"2025-12-01T17:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.843389 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.843416 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.843424 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.843439 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.843448 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:33Z","lastTransitionTime":"2025-12-01T17:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.945411 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.945469 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.945480 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.945500 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:33 crc kubenswrapper[4868]: I1201 17:26:33.945509 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:33Z","lastTransitionTime":"2025-12-01T17:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.047632 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.047666 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.047675 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.047689 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.047700 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:34Z","lastTransitionTime":"2025-12-01T17:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.150392 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.150451 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.150465 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.150492 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.150507 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:34Z","lastTransitionTime":"2025-12-01T17:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.171807 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.171827 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:34 crc kubenswrapper[4868]: E1201 17:26:34.171994 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:34 crc kubenswrapper[4868]: E1201 17:26:34.172029 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.253677 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.253716 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.253727 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.253745 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.253757 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:34Z","lastTransitionTime":"2025-12-01T17:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.356245 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.356292 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.356308 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.356325 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.356335 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:34Z","lastTransitionTime":"2025-12-01T17:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.459296 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.459345 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.459356 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.459381 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.459393 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:34Z","lastTransitionTime":"2025-12-01T17:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.562092 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.562129 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.562140 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.562158 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.562171 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:34Z","lastTransitionTime":"2025-12-01T17:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.664208 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.664696 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.664893 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.665057 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.665188 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:34Z","lastTransitionTime":"2025-12-01T17:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.768096 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.768640 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.768790 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.769008 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.769243 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:34Z","lastTransitionTime":"2025-12-01T17:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.873507 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.873563 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.873580 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.873604 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.873617 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:34Z","lastTransitionTime":"2025-12-01T17:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.976385 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.976430 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.976441 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.976460 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:34 crc kubenswrapper[4868]: I1201 17:26:34.976474 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:34Z","lastTransitionTime":"2025-12-01T17:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.079068 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.079106 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.079115 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.079138 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.079148 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:35Z","lastTransitionTime":"2025-12-01T17:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.171902 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.171937 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:35 crc kubenswrapper[4868]: E1201 17:26:35.172067 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:35 crc kubenswrapper[4868]: E1201 17:26:35.172212 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.181627 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.181670 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.181683 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.181701 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.181712 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:35Z","lastTransitionTime":"2025-12-01T17:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.284711 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.284786 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.284799 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.284819 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.284831 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:35Z","lastTransitionTime":"2025-12-01T17:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.387672 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.387721 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.387731 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.387748 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.387758 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:35Z","lastTransitionTime":"2025-12-01T17:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.490363 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.490407 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.490419 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.490436 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.490449 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:35Z","lastTransitionTime":"2025-12-01T17:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.592807 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.592849 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.592862 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.592881 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.592893 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:35Z","lastTransitionTime":"2025-12-01T17:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.695585 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.695626 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.695638 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.695657 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.695671 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:35Z","lastTransitionTime":"2025-12-01T17:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.797951 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.798001 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.798014 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.798038 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.798058 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:35Z","lastTransitionTime":"2025-12-01T17:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.900462 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.900516 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.900527 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.900546 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:35 crc kubenswrapper[4868]: I1201 17:26:35.900559 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:35Z","lastTransitionTime":"2025-12-01T17:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.002972 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.003008 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.003016 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.003031 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.003040 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:36Z","lastTransitionTime":"2025-12-01T17:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.105896 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.106013 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.106034 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.106061 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.106087 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:36Z","lastTransitionTime":"2025-12-01T17:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.171153 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.171171 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:36 crc kubenswrapper[4868]: E1201 17:26:36.171352 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:36 crc kubenswrapper[4868]: E1201 17:26:36.171601 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.209139 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.209181 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.209191 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.209209 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.209221 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:36Z","lastTransitionTime":"2025-12-01T17:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.311736 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.311791 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.311800 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.311817 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.311826 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:36Z","lastTransitionTime":"2025-12-01T17:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.414558 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.414623 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.414633 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.414648 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.414658 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:36Z","lastTransitionTime":"2025-12-01T17:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.517634 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.517689 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.517702 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.517722 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.517735 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:36Z","lastTransitionTime":"2025-12-01T17:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.620768 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.620827 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.620841 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.620862 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.620875 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:36Z","lastTransitionTime":"2025-12-01T17:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.724694 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.724734 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.724742 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.724757 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.724768 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:36Z","lastTransitionTime":"2025-12-01T17:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.827575 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.827627 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.827638 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.827655 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.827665 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:36Z","lastTransitionTime":"2025-12-01T17:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.930305 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.930362 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.930374 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.930392 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:36 crc kubenswrapper[4868]: I1201 17:26:36.930404 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:36Z","lastTransitionTime":"2025-12-01T17:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.033217 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.033265 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.033278 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.033295 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.033307 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:37Z","lastTransitionTime":"2025-12-01T17:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.135789 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.135840 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.135850 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.135868 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.135879 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:37Z","lastTransitionTime":"2025-12-01T17:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.171741 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.171739 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:37 crc kubenswrapper[4868]: E1201 17:26:37.171887 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:37 crc kubenswrapper[4868]: E1201 17:26:37.172009 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.237642 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.237685 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.237696 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.237712 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.237723 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:37Z","lastTransitionTime":"2025-12-01T17:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.340855 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.340971 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.340991 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.341012 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.341024 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:37Z","lastTransitionTime":"2025-12-01T17:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.443110 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.443141 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.443151 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.443165 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.443175 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:37Z","lastTransitionTime":"2025-12-01T17:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.545209 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.545238 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.545246 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.545260 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.545270 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:37Z","lastTransitionTime":"2025-12-01T17:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.647644 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.647706 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.647721 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.647740 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.647753 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:37Z","lastTransitionTime":"2025-12-01T17:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.749549 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.749585 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.749595 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.749608 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.749619 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:37Z","lastTransitionTime":"2025-12-01T17:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.851846 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.851886 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.851895 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.851911 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.851923 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:37Z","lastTransitionTime":"2025-12-01T17:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.954318 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.954360 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.954373 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.954390 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:37 crc kubenswrapper[4868]: I1201 17:26:37.954398 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:37Z","lastTransitionTime":"2025-12-01T17:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.056935 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.057029 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.057048 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.057082 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.057106 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:38Z","lastTransitionTime":"2025-12-01T17:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.159599 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.159630 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.159639 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.159654 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.159664 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:38Z","lastTransitionTime":"2025-12-01T17:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.171136 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:38 crc kubenswrapper[4868]: E1201 17:26:38.171277 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.171334 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:38 crc kubenswrapper[4868]: E1201 17:26:38.171808 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.172067 4868 scope.go:117] "RemoveContainer" containerID="d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb" Dec 01 17:26:38 crc kubenswrapper[4868]: E1201 17:26:38.172276 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-q9z6l_openshift-ovn-kubernetes(5baca705-90cc-4198-b641-214a2b915086)\"" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podUID="5baca705-90cc-4198-b641-214a2b915086" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.187252 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.201997 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.215155 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.228702 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.240027 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.252168 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.262291 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.262330 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.262342 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.262358 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.262368 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:38Z","lastTransitionTime":"2025-12-01T17:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.268280 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.282432 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.299625 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.318331 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b06e8b943027f83e6eaa247d071102f3f5feddf12d9a759692613820d468ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:09Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.556807 6333 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.556910 6333 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 17:26:09.557032 6333 factory.go:656] Stopping watch factory\\\\nI1201 17:26:09.557052 6333 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557341 6333 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 17:26:09.557758 6333 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 17:26:09.559974 6333 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1201 17:26:09.559996 6333 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1201 17:26:09.560056 6333 ovnkube.go:599] Stopped ovnkube\\\\nI1201 17:26:09.560088 6333 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1201 17:26:09.560163 6333 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:24Z\\\",\\\"message\\\":\\\"io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00784edfb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:mhc-mtrc,Protocol:TCP,Port:8444,TargetPort:{1 0 mhc-mtrc},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: controller,},ClusterIP:10.217.4.167,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.167],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1201 17:26:24.043188 6530 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.329119 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.338971 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.351258 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cc7d784-77de-4109-ba8c-192bb74290f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c84b10381473d660a8159f66c6c2f012e0d46e588ef7f19a0f6950edc94806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://120b3cade14d3c00f0a22c5f4db0c5500a669bb07a267b65bf7a386938467f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://865f0cda78b8e8ef0cd3d52fd8c4baa2e0c0f67e9d2d89752ac7a7f5e5df0a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.366544 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.367455 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.367508 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.367528 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.367549 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.367565 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:38Z","lastTransitionTime":"2025-12-01T17:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.380999 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.392242 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.404619 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.417958 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cc7d784-77de-4109-ba8c-192bb74290f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c84b10381473d660a8159f66c6c2f012e0d46e588ef7f19a0f6950edc94806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://120b3cade14d3c00f0a22c5f4db0c5500a669bb07a267b65bf7a386938467f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://865f0cda78b8e8ef0cd3d52fd8c4baa2e0c0f67e9d2d89752ac7a7f5e5df0a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.431656 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.447052 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.460302 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.470533 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.470578 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.470588 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.470607 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.470619 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:38Z","lastTransitionTime":"2025-12-01T17:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.472792 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.484094 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.494430 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.506789 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.522544 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.537630 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.548966 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.566450 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.572739 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.572781 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.572791 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.572808 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.572817 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:38Z","lastTransitionTime":"2025-12-01T17:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.582443 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.595084 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.609449 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.627128 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.645746 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:24Z\\\",\\\"message\\\":\\\"io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00784edfb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:mhc-mtrc,Protocol:TCP,Port:8444,TargetPort:{1 0 mhc-mtrc},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: controller,},ClusterIP:10.217.4.167,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.167],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1201 17:26:24.043188 6530 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-q9z6l_openshift-ovn-kubernetes(5baca705-90cc-4198-b641-214a2b915086)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:38Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.674583 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.674633 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.674644 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.674659 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.674671 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:38Z","lastTransitionTime":"2025-12-01T17:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.778052 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.778205 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.778235 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.778281 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.778303 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:38Z","lastTransitionTime":"2025-12-01T17:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.881010 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.881052 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.881064 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.881082 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.881096 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:38Z","lastTransitionTime":"2025-12-01T17:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.983316 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.983800 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.983875 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.983965 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:38 crc kubenswrapper[4868]: I1201 17:26:38.984025 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:38Z","lastTransitionTime":"2025-12-01T17:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.086047 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.086088 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.086098 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.086114 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.086142 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:39Z","lastTransitionTime":"2025-12-01T17:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.170913 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.170986 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:39 crc kubenswrapper[4868]: E1201 17:26:39.171071 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:39 crc kubenswrapper[4868]: E1201 17:26:39.171129 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.188664 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.188700 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.188712 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.188728 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.188740 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:39Z","lastTransitionTime":"2025-12-01T17:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.291167 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.291219 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.291229 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.291247 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.291260 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:39Z","lastTransitionTime":"2025-12-01T17:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.394563 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.394619 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.394642 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.394667 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.394679 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:39Z","lastTransitionTime":"2025-12-01T17:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.497375 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.497404 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.497412 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.497427 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.497437 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:39Z","lastTransitionTime":"2025-12-01T17:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.600337 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.600370 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.600378 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.600392 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.600400 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:39Z","lastTransitionTime":"2025-12-01T17:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.702665 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.702709 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.702719 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.702735 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.702747 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:39Z","lastTransitionTime":"2025-12-01T17:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.805066 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.805143 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.805157 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.805179 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.805194 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:39Z","lastTransitionTime":"2025-12-01T17:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.908120 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.908164 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.908172 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.908187 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:39 crc kubenswrapper[4868]: I1201 17:26:39.908198 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:39Z","lastTransitionTime":"2025-12-01T17:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.010756 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.010797 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.010808 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.010823 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.010833 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:40Z","lastTransitionTime":"2025-12-01T17:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.113372 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.113422 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.113433 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.113451 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.113465 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:40Z","lastTransitionTime":"2025-12-01T17:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.171329 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.171382 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:40 crc kubenswrapper[4868]: E1201 17:26:40.171481 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:40 crc kubenswrapper[4868]: E1201 17:26:40.171619 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.215541 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.215580 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.215589 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.215604 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.215616 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:40Z","lastTransitionTime":"2025-12-01T17:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.318345 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.318389 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.318402 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.318421 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.318437 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:40Z","lastTransitionTime":"2025-12-01T17:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.420579 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.420620 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.420630 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.420646 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.420656 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:40Z","lastTransitionTime":"2025-12-01T17:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.523748 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.523791 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.523804 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.523822 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.523834 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:40Z","lastTransitionTime":"2025-12-01T17:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.626473 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.626532 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.626544 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.626565 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.626579 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:40Z","lastTransitionTime":"2025-12-01T17:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.676831 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs\") pod \"network-metrics-daemon-nxfkx\" (UID: \"4afd9a0c-e8c7-42ec-9771-036c3441cbce\") " pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:40 crc kubenswrapper[4868]: E1201 17:26:40.677027 4868 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:26:40 crc kubenswrapper[4868]: E1201 17:26:40.677107 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs podName:4afd9a0c-e8c7-42ec-9771-036c3441cbce nodeName:}" failed. No retries permitted until 2025-12-01 17:27:12.677087279 +0000 UTC m=+105.048197690 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs") pod "network-metrics-daemon-nxfkx" (UID: "4afd9a0c-e8c7-42ec-9771-036c3441cbce") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.729803 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.729868 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.729886 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.729913 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.729928 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:40Z","lastTransitionTime":"2025-12-01T17:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.832759 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.832812 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.832825 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.832845 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.832858 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:40Z","lastTransitionTime":"2025-12-01T17:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.936006 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.936069 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.936081 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.936103 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:40 crc kubenswrapper[4868]: I1201 17:26:40.936116 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:40Z","lastTransitionTime":"2025-12-01T17:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.038828 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.038883 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.038895 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.038912 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.038921 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:41Z","lastTransitionTime":"2025-12-01T17:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.141350 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.141391 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.141400 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.141416 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.141426 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:41Z","lastTransitionTime":"2025-12-01T17:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.171565 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.171565 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:41 crc kubenswrapper[4868]: E1201 17:26:41.171690 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:41 crc kubenswrapper[4868]: E1201 17:26:41.171762 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.244833 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.244894 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.244911 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.244932 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.244971 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:41Z","lastTransitionTime":"2025-12-01T17:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.347314 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.347350 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.347359 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.347376 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.347385 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:41Z","lastTransitionTime":"2025-12-01T17:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.449507 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.449554 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.449567 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.449641 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.449655 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:41Z","lastTransitionTime":"2025-12-01T17:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.552854 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.552929 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.552954 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.552968 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.552977 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:41Z","lastTransitionTime":"2025-12-01T17:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.655347 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.655401 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.655414 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.655433 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.655444 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:41Z","lastTransitionTime":"2025-12-01T17:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.757402 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.757436 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.757445 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.757460 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.757469 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:41Z","lastTransitionTime":"2025-12-01T17:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.859506 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.859537 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.859545 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.859559 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.859568 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:41Z","lastTransitionTime":"2025-12-01T17:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.961808 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.961862 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.961873 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.961889 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:41 crc kubenswrapper[4868]: I1201 17:26:41.961897 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:41Z","lastTransitionTime":"2025-12-01T17:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.064472 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.064511 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.064524 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.064542 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.064556 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:42Z","lastTransitionTime":"2025-12-01T17:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.167020 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.167055 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.167063 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.167095 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.167107 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:42Z","lastTransitionTime":"2025-12-01T17:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.174094 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:42 crc kubenswrapper[4868]: E1201 17:26:42.174311 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.174359 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:42 crc kubenswrapper[4868]: E1201 17:26:42.174576 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.269623 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.269658 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.269668 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.269686 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.269696 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:42Z","lastTransitionTime":"2025-12-01T17:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.371856 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.371896 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.371906 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.371928 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.371954 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:42Z","lastTransitionTime":"2025-12-01T17:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.474238 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.474284 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.474295 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.474311 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.474322 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:42Z","lastTransitionTime":"2025-12-01T17:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.576897 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.576932 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.576959 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.576976 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.576986 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:42Z","lastTransitionTime":"2025-12-01T17:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.679337 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.679379 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.679389 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.679406 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.679420 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:42Z","lastTransitionTime":"2025-12-01T17:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.782254 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.782298 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.782311 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.782331 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.782343 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:42Z","lastTransitionTime":"2025-12-01T17:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.884979 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.885033 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.885047 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.885068 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.885083 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:42Z","lastTransitionTime":"2025-12-01T17:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.988031 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.988088 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.988103 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.988123 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:42 crc kubenswrapper[4868]: I1201 17:26:42.988137 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:42Z","lastTransitionTime":"2025-12-01T17:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.090868 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.090916 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.090927 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.090962 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.090974 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:43Z","lastTransitionTime":"2025-12-01T17:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.169899 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.170067 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.170078 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.170095 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.170104 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:43Z","lastTransitionTime":"2025-12-01T17:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.171267 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.171339 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:43 crc kubenswrapper[4868]: E1201 17:26:43.171401 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:43 crc kubenswrapper[4868]: E1201 17:26:43.171592 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:43 crc kubenswrapper[4868]: E1201 17:26:43.187360 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:43Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.191770 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.191829 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.191839 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.191858 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.191868 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:43Z","lastTransitionTime":"2025-12-01T17:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:43 crc kubenswrapper[4868]: E1201 17:26:43.204643 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:43Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.208787 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.208830 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.208847 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.208869 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.208883 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:43Z","lastTransitionTime":"2025-12-01T17:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:43 crc kubenswrapper[4868]: E1201 17:26:43.223853 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:43Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.229317 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.229384 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.229396 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.229418 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.229430 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:43Z","lastTransitionTime":"2025-12-01T17:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:43 crc kubenswrapper[4868]: E1201 17:26:43.243484 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:43Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.248291 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.248349 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.248363 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.248383 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.248397 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:43Z","lastTransitionTime":"2025-12-01T17:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:43 crc kubenswrapper[4868]: E1201 17:26:43.264032 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:43Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:43 crc kubenswrapper[4868]: E1201 17:26:43.264172 4868 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.265874 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.265902 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.265912 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.265930 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.265958 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:43Z","lastTransitionTime":"2025-12-01T17:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.368435 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.368483 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.368531 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.368554 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.368567 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:43Z","lastTransitionTime":"2025-12-01T17:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.471389 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.471426 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.471434 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.471449 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.471458 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:43Z","lastTransitionTime":"2025-12-01T17:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.573969 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.574013 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.574038 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.574054 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.574062 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:43Z","lastTransitionTime":"2025-12-01T17:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.675783 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.675823 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.675833 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.675850 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.675861 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:43Z","lastTransitionTime":"2025-12-01T17:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.778828 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.778886 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.778899 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.778919 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.778932 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:43Z","lastTransitionTime":"2025-12-01T17:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.881306 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.881429 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.881442 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.881462 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.881471 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:43Z","lastTransitionTime":"2025-12-01T17:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.983859 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.983908 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.983918 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.983934 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:43 crc kubenswrapper[4868]: I1201 17:26:43.983959 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:43Z","lastTransitionTime":"2025-12-01T17:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.086802 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.086870 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.086893 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.086926 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.086980 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:44Z","lastTransitionTime":"2025-12-01T17:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.171917 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.172027 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:44 crc kubenswrapper[4868]: E1201 17:26:44.172081 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:44 crc kubenswrapper[4868]: E1201 17:26:44.172153 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.189161 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.189194 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.189203 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.189217 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.189227 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:44Z","lastTransitionTime":"2025-12-01T17:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.291935 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.292024 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.292043 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.292063 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.292075 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:44Z","lastTransitionTime":"2025-12-01T17:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.394084 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.394133 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.394153 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.394469 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.394504 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:44Z","lastTransitionTime":"2025-12-01T17:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.496682 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.496739 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.496753 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.496780 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.496792 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:44Z","lastTransitionTime":"2025-12-01T17:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.599603 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.599653 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.599667 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.599687 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.599697 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:44Z","lastTransitionTime":"2025-12-01T17:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.701867 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.701915 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.701927 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.701960 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.701970 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:44Z","lastTransitionTime":"2025-12-01T17:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.804010 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.804065 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.804077 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.804095 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.804107 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:44Z","lastTransitionTime":"2025-12-01T17:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.906444 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.906501 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.906513 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.906530 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:44 crc kubenswrapper[4868]: I1201 17:26:44.906539 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:44Z","lastTransitionTime":"2025-12-01T17:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.008929 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.008994 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.009003 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.009019 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.009031 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:45Z","lastTransitionTime":"2025-12-01T17:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.111267 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.111307 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.111317 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.111333 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.111342 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:45Z","lastTransitionTime":"2025-12-01T17:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.171074 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.171111 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:45 crc kubenswrapper[4868]: E1201 17:26:45.171233 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:45 crc kubenswrapper[4868]: E1201 17:26:45.171341 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.214175 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.214223 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.214234 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.214253 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.214266 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:45Z","lastTransitionTime":"2025-12-01T17:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.316488 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.316537 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.316549 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.316567 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.316580 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:45Z","lastTransitionTime":"2025-12-01T17:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.419119 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.419166 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.419176 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.419196 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.419208 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:45Z","lastTransitionTime":"2025-12-01T17:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.521251 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.521296 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.521307 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.521323 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.521335 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:45Z","lastTransitionTime":"2025-12-01T17:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.624044 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.624107 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.624129 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.624154 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.624173 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:45Z","lastTransitionTime":"2025-12-01T17:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.687702 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9tlw_8ce688b4-36b1-48d4-a6d1-8aec723125c4/kube-multus/0.log" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.687781 4868 generic.go:334] "Generic (PLEG): container finished" podID="8ce688b4-36b1-48d4-a6d1-8aec723125c4" containerID="b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58" exitCode=1 Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.687830 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9tlw" event={"ID":"8ce688b4-36b1-48d4-a6d1-8aec723125c4","Type":"ContainerDied","Data":"b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58"} Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.688427 4868 scope.go:117] "RemoveContainer" containerID="b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.706623 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:45Z\\\",\\\"message\\\":\\\"2025-12-01T17:26:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0\\\\n2025-12-01T17:26:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0 to /host/opt/cni/bin/\\\\n2025-12-01T17:26:00Z [verbose] multus-daemon started\\\\n2025-12-01T17:26:00Z [verbose] Readiness Indicator file check\\\\n2025-12-01T17:26:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.725138 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.729080 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.729112 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.729127 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.729145 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.729155 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:45Z","lastTransitionTime":"2025-12-01T17:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.744804 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.759144 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.775111 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.795881 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:24Z\\\",\\\"message\\\":\\\"io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00784edfb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:mhc-mtrc,Protocol:TCP,Port:8444,TargetPort:{1 0 mhc-mtrc},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: controller,},ClusterIP:10.217.4.167,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.167],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1201 17:26:24.043188 6530 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-q9z6l_openshift-ovn-kubernetes(5baca705-90cc-4198-b641-214a2b915086)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.809246 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.821081 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.831722 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.831749 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.831758 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.831774 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.831784 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:45Z","lastTransitionTime":"2025-12-01T17:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.832484 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.844839 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.860916 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.874174 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.885529 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.898063 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cc7d784-77de-4109-ba8c-192bb74290f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c84b10381473d660a8159f66c6c2f012e0d46e588ef7f19a0f6950edc94806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://120b3cade14d3c00f0a22c5f4db0c5500a669bb07a267b65bf7a386938467f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://865f0cda78b8e8ef0cd3d52fd8c4baa2e0c0f67e9d2d89752ac7a7f5e5df0a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.913016 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.926887 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.934515 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.934557 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.934570 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.934589 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.934601 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:45Z","lastTransitionTime":"2025-12-01T17:26:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:45 crc kubenswrapper[4868]: I1201 17:26:45.941119 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:45Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.036958 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.037000 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.037012 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.037028 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.037074 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:46Z","lastTransitionTime":"2025-12-01T17:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.139633 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.139776 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.139795 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.139811 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.139821 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:46Z","lastTransitionTime":"2025-12-01T17:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.171213 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.171216 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:46 crc kubenswrapper[4868]: E1201 17:26:46.171422 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:46 crc kubenswrapper[4868]: E1201 17:26:46.171545 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.242668 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.242711 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.242724 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.242742 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.242752 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:46Z","lastTransitionTime":"2025-12-01T17:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.345515 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.345569 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.345578 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.345595 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.345607 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:46Z","lastTransitionTime":"2025-12-01T17:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.447554 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.447623 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.447640 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.447672 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.447687 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:46Z","lastTransitionTime":"2025-12-01T17:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.550612 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.550652 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.550662 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.550675 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.550685 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:46Z","lastTransitionTime":"2025-12-01T17:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.653236 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.653279 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.653289 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.653307 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.653317 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:46Z","lastTransitionTime":"2025-12-01T17:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.692930 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9tlw_8ce688b4-36b1-48d4-a6d1-8aec723125c4/kube-multus/0.log" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.693009 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9tlw" event={"ID":"8ce688b4-36b1-48d4-a6d1-8aec723125c4","Type":"ContainerStarted","Data":"3d8b1e3f6f1f6f8ed2844b1e533905130d255430f7031699ce02dfa6a3ac16a5"} Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.706399 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d8b1e3f6f1f6f8ed2844b1e533905130d255430f7031699ce02dfa6a3ac16a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:45Z\\\",\\\"message\\\":\\\"2025-12-01T17:26:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0\\\\n2025-12-01T17:26:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0 to /host/opt/cni/bin/\\\\n2025-12-01T17:26:00Z [verbose] multus-daemon started\\\\n2025-12-01T17:26:00Z [verbose] Readiness Indicator file check\\\\n2025-12-01T17:26:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.720170 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.736213 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.753655 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:24Z\\\",\\\"message\\\":\\\"io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00784edfb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:mhc-mtrc,Protocol:TCP,Port:8444,TargetPort:{1 0 mhc-mtrc},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: controller,},ClusterIP:10.217.4.167,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.167],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1201 17:26:24.043188 6530 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-q9z6l_openshift-ovn-kubernetes(5baca705-90cc-4198-b641-214a2b915086)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.755367 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.755395 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.755404 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.755419 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.755430 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:46Z","lastTransitionTime":"2025-12-01T17:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.764520 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.776760 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.789282 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.802182 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.812085 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.823393 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.840388 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.854885 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cc7d784-77de-4109-ba8c-192bb74290f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c84b10381473d660a8159f66c6c2f012e0d46e588ef7f19a0f6950edc94806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://120b3cade14d3c00f0a22c5f4db0c5500a669bb07a267b65bf7a386938467f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://865f0cda78b8e8ef0cd3d52fd8c4baa2e0c0f67e9d2d89752ac7a7f5e5df0a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.858734 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.858796 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.858805 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.858824 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.858834 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:46Z","lastTransitionTime":"2025-12-01T17:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.870418 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.883506 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.894360 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.909382 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.923054 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:46Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.961359 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.961397 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.961406 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.961421 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:46 crc kubenswrapper[4868]: I1201 17:26:46.961431 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:46Z","lastTransitionTime":"2025-12-01T17:26:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.064086 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.064132 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.064140 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.064156 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.064165 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:47Z","lastTransitionTime":"2025-12-01T17:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.167287 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.167349 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.167363 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.167388 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.167405 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:47Z","lastTransitionTime":"2025-12-01T17:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.171647 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:47 crc kubenswrapper[4868]: E1201 17:26:47.171818 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.171917 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:47 crc kubenswrapper[4868]: E1201 17:26:47.172106 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.269772 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.269816 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.269825 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.269842 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.269851 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:47Z","lastTransitionTime":"2025-12-01T17:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.371923 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.371976 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.371987 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.372008 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.372021 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:47Z","lastTransitionTime":"2025-12-01T17:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.474470 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.474507 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.474516 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.474531 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.474542 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:47Z","lastTransitionTime":"2025-12-01T17:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.577144 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.577203 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.577213 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.577232 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.577244 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:47Z","lastTransitionTime":"2025-12-01T17:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.679517 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.679576 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.679592 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.679615 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.679631 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:47Z","lastTransitionTime":"2025-12-01T17:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.781644 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.781687 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.781700 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.781717 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.781729 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:47Z","lastTransitionTime":"2025-12-01T17:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.884003 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.884047 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.884057 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.884075 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.884087 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:47Z","lastTransitionTime":"2025-12-01T17:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.986761 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.986825 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.986834 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.986857 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:47 crc kubenswrapper[4868]: I1201 17:26:47.986867 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:47Z","lastTransitionTime":"2025-12-01T17:26:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.089325 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.089365 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.089376 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.089393 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.089405 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:48Z","lastTransitionTime":"2025-12-01T17:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.171836 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.171880 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:48 crc kubenswrapper[4868]: E1201 17:26:48.172027 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:48 crc kubenswrapper[4868]: E1201 17:26:48.172183 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.190141 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d8b1e3f6f1f6f8ed2844b1e533905130d255430f7031699ce02dfa6a3ac16a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:45Z\\\",\\\"message\\\":\\\"2025-12-01T17:26:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0\\\\n2025-12-01T17:26:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0 to /host/opt/cni/bin/\\\\n2025-12-01T17:26:00Z [verbose] multus-daemon started\\\\n2025-12-01T17:26:00Z [verbose] Readiness Indicator file check\\\\n2025-12-01T17:26:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.191245 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.191274 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.191289 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.191307 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.191319 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:48Z","lastTransitionTime":"2025-12-01T17:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.202777 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.216893 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.232901 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.246599 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.262716 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.286120 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:24Z\\\",\\\"message\\\":\\\"io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00784edfb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:mhc-mtrc,Protocol:TCP,Port:8444,TargetPort:{1 0 mhc-mtrc},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: controller,},ClusterIP:10.217.4.167,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.167],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1201 17:26:24.043188 6530 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-q9z6l_openshift-ovn-kubernetes(5baca705-90cc-4198-b641-214a2b915086)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.293723 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.293751 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.293759 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.293774 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.293784 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:48Z","lastTransitionTime":"2025-12-01T17:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.299023 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.311857 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.323254 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cc7d784-77de-4109-ba8c-192bb74290f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c84b10381473d660a8159f66c6c2f012e0d46e588ef7f19a0f6950edc94806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://120b3cade14d3c00f0a22c5f4db0c5500a669bb07a267b65bf7a386938467f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://865f0cda78b8e8ef0cd3d52fd8c4baa2e0c0f67e9d2d89752ac7a7f5e5df0a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.336807 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.352124 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.364893 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.379535 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.395721 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.396683 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.396707 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.396717 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.396750 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.396759 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:48Z","lastTransitionTime":"2025-12-01T17:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.409541 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.422293 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:48Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.499427 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.499465 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.499476 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.499491 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.499500 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:48Z","lastTransitionTime":"2025-12-01T17:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.601220 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.601265 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.601276 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.601293 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.601304 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:48Z","lastTransitionTime":"2025-12-01T17:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.703129 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.703213 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.703225 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.703240 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.703251 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:48Z","lastTransitionTime":"2025-12-01T17:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.805778 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.805808 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.805818 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.805831 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.805841 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:48Z","lastTransitionTime":"2025-12-01T17:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.910742 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.911395 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.911505 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.911579 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:48 crc kubenswrapper[4868]: I1201 17:26:48.911637 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:48Z","lastTransitionTime":"2025-12-01T17:26:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.014602 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.014916 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.015024 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.015108 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.015176 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:49Z","lastTransitionTime":"2025-12-01T17:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.117352 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.117393 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.117401 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.117418 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.117429 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:49Z","lastTransitionTime":"2025-12-01T17:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.171062 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.171136 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:49 crc kubenswrapper[4868]: E1201 17:26:49.171497 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:49 crc kubenswrapper[4868]: E1201 17:26:49.171557 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.219657 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.219695 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.219705 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.219720 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.219729 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:49Z","lastTransitionTime":"2025-12-01T17:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.323879 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.323917 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.323930 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.323963 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.323977 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:49Z","lastTransitionTime":"2025-12-01T17:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.426141 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.426181 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.426191 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.426207 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.426217 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:49Z","lastTransitionTime":"2025-12-01T17:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.528716 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.528773 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.528785 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.528806 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.528820 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:49Z","lastTransitionTime":"2025-12-01T17:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.631763 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.631809 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.631819 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.631836 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.631846 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:49Z","lastTransitionTime":"2025-12-01T17:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.734628 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.734680 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.734694 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.734712 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.734724 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:49Z","lastTransitionTime":"2025-12-01T17:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.836912 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.836985 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.837001 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.837019 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.837031 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:49Z","lastTransitionTime":"2025-12-01T17:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.939765 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.939810 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.939821 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.939838 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:49 crc kubenswrapper[4868]: I1201 17:26:49.939847 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:49Z","lastTransitionTime":"2025-12-01T17:26:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.043001 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.043049 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.043058 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.043077 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.043086 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:50Z","lastTransitionTime":"2025-12-01T17:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.145801 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.145847 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.145859 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.145876 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.145886 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:50Z","lastTransitionTime":"2025-12-01T17:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.171352 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.171432 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:50 crc kubenswrapper[4868]: E1201 17:26:50.171631 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:50 crc kubenswrapper[4868]: E1201 17:26:50.171887 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.248829 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.248867 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.248879 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.248897 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.248907 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:50Z","lastTransitionTime":"2025-12-01T17:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.352029 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.352077 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.352093 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.352177 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.352194 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:50Z","lastTransitionTime":"2025-12-01T17:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.455188 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.455506 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.455515 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.455531 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.455542 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:50Z","lastTransitionTime":"2025-12-01T17:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.558237 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.558277 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.558288 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.558304 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.558314 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:50Z","lastTransitionTime":"2025-12-01T17:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.661428 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.661473 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.661486 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.661507 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.661517 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:50Z","lastTransitionTime":"2025-12-01T17:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.764339 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.764383 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.764392 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.764426 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.764437 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:50Z","lastTransitionTime":"2025-12-01T17:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.866442 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.866493 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.866505 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.866523 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.866533 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:50Z","lastTransitionTime":"2025-12-01T17:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.968516 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.968548 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.968557 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.968570 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:50 crc kubenswrapper[4868]: I1201 17:26:50.968580 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:50Z","lastTransitionTime":"2025-12-01T17:26:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.070342 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.070379 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.070389 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.070405 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.070416 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:51Z","lastTransitionTime":"2025-12-01T17:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.084404 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.084502 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:51 crc kubenswrapper[4868]: E1201 17:26:51.084569 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:27:55.0845437 +0000 UTC m=+147.455654111 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.084603 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:51 crc kubenswrapper[4868]: E1201 17:26:51.084613 4868 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:26:51 crc kubenswrapper[4868]: E1201 17:26:51.084655 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:27:55.084646532 +0000 UTC m=+147.455756943 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 17:26:51 crc kubenswrapper[4868]: E1201 17:26:51.084669 4868 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:26:51 crc kubenswrapper[4868]: E1201 17:26:51.084701 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 17:27:55.084692364 +0000 UTC m=+147.455802775 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.171780 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.171862 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:51 crc kubenswrapper[4868]: E1201 17:26:51.171918 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:51 crc kubenswrapper[4868]: E1201 17:26:51.171966 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.173190 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.173227 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.173243 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.173257 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.173268 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:51Z","lastTransitionTime":"2025-12-01T17:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.185733 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.185783 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:51 crc kubenswrapper[4868]: E1201 17:26:51.185900 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:26:51 crc kubenswrapper[4868]: E1201 17:26:51.185915 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:26:51 crc kubenswrapper[4868]: E1201 17:26:51.185925 4868 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:26:51 crc kubenswrapper[4868]: E1201 17:26:51.186022 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 17:27:55.186004505 +0000 UTC m=+147.557114916 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:26:51 crc kubenswrapper[4868]: E1201 17:26:51.186017 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 17:26:51 crc kubenswrapper[4868]: E1201 17:26:51.186061 4868 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 17:26:51 crc kubenswrapper[4868]: E1201 17:26:51.186075 4868 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:26:51 crc kubenswrapper[4868]: E1201 17:26:51.186145 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 17:27:55.186121948 +0000 UTC m=+147.557232359 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.276334 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.276387 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.276398 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.276414 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.276425 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:51Z","lastTransitionTime":"2025-12-01T17:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.378720 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.378763 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.378772 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.378788 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.378800 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:51Z","lastTransitionTime":"2025-12-01T17:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.482135 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.482178 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.482191 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.482210 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.482222 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:51Z","lastTransitionTime":"2025-12-01T17:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.584776 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.584829 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.584843 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.584861 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.584873 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:51Z","lastTransitionTime":"2025-12-01T17:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.687747 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.687785 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.687793 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.687813 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.687822 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:51Z","lastTransitionTime":"2025-12-01T17:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.790379 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.790425 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.790436 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.790452 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.790461 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:51Z","lastTransitionTime":"2025-12-01T17:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.892692 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.892728 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.892737 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.892754 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.892764 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:51Z","lastTransitionTime":"2025-12-01T17:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.995901 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.995962 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.995973 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.995993 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:51 crc kubenswrapper[4868]: I1201 17:26:51.996005 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:51Z","lastTransitionTime":"2025-12-01T17:26:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.098522 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.098559 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.098571 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.098589 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.098601 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:52Z","lastTransitionTime":"2025-12-01T17:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.171266 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.171298 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:52 crc kubenswrapper[4868]: E1201 17:26:52.171438 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:52 crc kubenswrapper[4868]: E1201 17:26:52.171859 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.172309 4868 scope.go:117] "RemoveContainer" containerID="d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.200450 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.200514 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.200524 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.200542 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.200555 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:52Z","lastTransitionTime":"2025-12-01T17:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.305363 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.305404 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.305414 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.305428 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.305440 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:52Z","lastTransitionTime":"2025-12-01T17:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.392932 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.407799 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.407833 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.407843 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.407857 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.407868 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:52Z","lastTransitionTime":"2025-12-01T17:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.510434 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.510480 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.510490 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.510508 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.510520 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:52Z","lastTransitionTime":"2025-12-01T17:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.613053 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.613108 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.613118 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.613135 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.613144 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:52Z","lastTransitionTime":"2025-12-01T17:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.714775 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovnkube-controller/2.log" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.714848 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.714887 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.714897 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.714917 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.714931 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:52Z","lastTransitionTime":"2025-12-01T17:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.718544 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerStarted","Data":"fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c"} Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.719395 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.735584 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.756493 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:24Z\\\",\\\"message\\\":\\\"io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00784edfb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:mhc-mtrc,Protocol:TCP,Port:8444,TargetPort:{1 0 mhc-mtrc},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: controller,},ClusterIP:10.217.4.167,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.167],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1201 17:26:24.043188 6530 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.768149 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.778826 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.789470 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.801303 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.812314 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.818183 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.818228 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.818239 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.818256 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.818266 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:52Z","lastTransitionTime":"2025-12-01T17:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.825034 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.837754 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.850107 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cc7d784-77de-4109-ba8c-192bb74290f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c84b10381473d660a8159f66c6c2f012e0d46e588ef7f19a0f6950edc94806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://120b3cade14d3c00f0a22c5f4db0c5500a669bb07a267b65bf7a386938467f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://865f0cda78b8e8ef0cd3d52fd8c4baa2e0c0f67e9d2d89752ac7a7f5e5df0a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.863979 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.876866 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.893021 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.908815 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.920853 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.920900 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.920914 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.920933 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.920966 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:52Z","lastTransitionTime":"2025-12-01T17:26:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.922563 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.936589 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d8b1e3f6f1f6f8ed2844b1e533905130d255430f7031699ce02dfa6a3ac16a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:45Z\\\",\\\"message\\\":\\\"2025-12-01T17:26:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0\\\\n2025-12-01T17:26:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0 to /host/opt/cni/bin/\\\\n2025-12-01T17:26:00Z [verbose] multus-daemon started\\\\n2025-12-01T17:26:00Z [verbose] Readiness Indicator file check\\\\n2025-12-01T17:26:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:52 crc kubenswrapper[4868]: I1201 17:26:52.948653 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:52Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.024178 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.024224 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.024237 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.024259 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.024275 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:53Z","lastTransitionTime":"2025-12-01T17:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.126339 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.126374 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.126386 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.126403 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.126414 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:53Z","lastTransitionTime":"2025-12-01T17:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.171340 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.171393 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:53 crc kubenswrapper[4868]: E1201 17:26:53.171638 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:53 crc kubenswrapper[4868]: E1201 17:26:53.171855 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.229123 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.229167 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.229178 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.229196 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.229206 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:53Z","lastTransitionTime":"2025-12-01T17:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.332034 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.332084 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.332095 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.332117 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.332130 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:53Z","lastTransitionTime":"2025-12-01T17:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.406662 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.406713 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.406725 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.406745 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.406759 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:53Z","lastTransitionTime":"2025-12-01T17:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:53 crc kubenswrapper[4868]: E1201 17:26:53.420394 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:53Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.424461 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.424509 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.424520 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.424537 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.424549 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:53Z","lastTransitionTime":"2025-12-01T17:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:53 crc kubenswrapper[4868]: E1201 17:26:53.438492 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:53Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.442964 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.443000 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.443009 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.443024 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.443035 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:53Z","lastTransitionTime":"2025-12-01T17:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:53 crc kubenswrapper[4868]: E1201 17:26:53.454411 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:53Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.457583 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.457800 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.457810 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.457831 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.457841 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:53Z","lastTransitionTime":"2025-12-01T17:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:53 crc kubenswrapper[4868]: E1201 17:26:53.482360 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:53Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.486109 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.486139 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.486148 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.486163 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.486174 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:53Z","lastTransitionTime":"2025-12-01T17:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:53 crc kubenswrapper[4868]: E1201 17:26:53.497776 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:53Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:53 crc kubenswrapper[4868]: E1201 17:26:53.497899 4868 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.499773 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.499826 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.499838 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.499857 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.499868 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:53Z","lastTransitionTime":"2025-12-01T17:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.602309 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.602357 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.602367 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.602383 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.602394 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:53Z","lastTransitionTime":"2025-12-01T17:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.705454 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.705500 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.705510 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.705529 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.705542 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:53Z","lastTransitionTime":"2025-12-01T17:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.807590 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.807643 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.807657 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.807679 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.807692 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:53Z","lastTransitionTime":"2025-12-01T17:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.909649 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.909698 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.909711 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.909728 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:53 crc kubenswrapper[4868]: I1201 17:26:53.909741 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:53Z","lastTransitionTime":"2025-12-01T17:26:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.012596 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.012644 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.012658 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.012677 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.012686 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:54Z","lastTransitionTime":"2025-12-01T17:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.115493 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.115550 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.115565 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.115589 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.115603 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:54Z","lastTransitionTime":"2025-12-01T17:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.170987 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:54 crc kubenswrapper[4868]: E1201 17:26:54.171175 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.171283 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:54 crc kubenswrapper[4868]: E1201 17:26:54.171454 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.218365 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.218413 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.218428 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.218454 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.218473 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:54Z","lastTransitionTime":"2025-12-01T17:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.321244 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.321292 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.321304 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.321324 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.321336 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:54Z","lastTransitionTime":"2025-12-01T17:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.424916 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.424977 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.424989 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.425009 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.425022 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:54Z","lastTransitionTime":"2025-12-01T17:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.528829 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.528896 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.528908 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.528925 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.528936 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:54Z","lastTransitionTime":"2025-12-01T17:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.631729 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.631787 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.631799 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.631819 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.631831 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:54Z","lastTransitionTime":"2025-12-01T17:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.735037 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.735073 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.735082 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.735097 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.735107 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:54Z","lastTransitionTime":"2025-12-01T17:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.837843 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.837881 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.837889 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.837905 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.837916 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:54Z","lastTransitionTime":"2025-12-01T17:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.941340 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.941395 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.941407 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.941427 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:54 crc kubenswrapper[4868]: I1201 17:26:54.941441 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:54Z","lastTransitionTime":"2025-12-01T17:26:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.044612 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.044686 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.044704 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.044732 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.044750 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:55Z","lastTransitionTime":"2025-12-01T17:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.147898 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.148028 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.148051 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.148081 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.148100 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:55Z","lastTransitionTime":"2025-12-01T17:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.171141 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.171192 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:55 crc kubenswrapper[4868]: E1201 17:26:55.171487 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:55 crc kubenswrapper[4868]: E1201 17:26:55.171631 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.251538 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.251617 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.251642 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.251671 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.251690 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:55Z","lastTransitionTime":"2025-12-01T17:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.354562 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.354612 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.354622 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.354638 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.354650 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:55Z","lastTransitionTime":"2025-12-01T17:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.456777 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.456812 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.456820 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.456836 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.456849 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:55Z","lastTransitionTime":"2025-12-01T17:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.559389 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.559445 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.559455 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.559471 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.559482 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:55Z","lastTransitionTime":"2025-12-01T17:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.662652 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.662710 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.662722 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.662741 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.662754 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:55Z","lastTransitionTime":"2025-12-01T17:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.730779 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovnkube-controller/3.log" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.732052 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovnkube-controller/2.log" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.740782 4868 generic.go:334] "Generic (PLEG): container finished" podID="5baca705-90cc-4198-b641-214a2b915086" containerID="fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c" exitCode=1 Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.740872 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerDied","Data":"fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c"} Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.740971 4868 scope.go:117] "RemoveContainer" containerID="d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.742886 4868 scope.go:117] "RemoveContainer" containerID="fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c" Dec 01 17:26:55 crc kubenswrapper[4868]: E1201 17:26:55.743245 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-q9z6l_openshift-ovn-kubernetes(5baca705-90cc-4198-b641-214a2b915086)\"" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podUID="5baca705-90cc-4198-b641-214a2b915086" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.758585 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.765262 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.765302 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.765314 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.765334 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.765346 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:55Z","lastTransitionTime":"2025-12-01T17:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.772594 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.783672 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.797291 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d8b1e3f6f1f6f8ed2844b1e533905130d255430f7031699ce02dfa6a3ac16a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:45Z\\\",\\\"message\\\":\\\"2025-12-01T17:26:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0\\\\n2025-12-01T17:26:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0 to /host/opt/cni/bin/\\\\n2025-12-01T17:26:00Z [verbose] multus-daemon started\\\\n2025-12-01T17:26:00Z [verbose] Readiness Indicator file check\\\\n2025-12-01T17:26:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.812846 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.831540 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.842778 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.861974 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.868468 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.868505 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.868516 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.868538 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.868551 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:55Z","lastTransitionTime":"2025-12-01T17:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.885090 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:24Z\\\",\\\"message\\\":\\\"io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00784edfb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:mhc-mtrc,Protocol:TCP,Port:8444,TargetPort:{1 0 mhc-mtrc},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: controller,},ClusterIP:10.217.4.167,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.167],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1201 17:26:24.043188 6530 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:55Z\\\",\\\"message\\\":\\\"t-ovn-kubernetes/ovnkube-node-q9z6l\\\\nI1201 17:26:55.063250 6930 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1201 17:26:55.063211 6930 services_controller.go:356] Processing sync for service default/kubernetes for network=default\\\\nI1201 17:26:55.063161 6930 services_controller.go:443] Built service openshift-operator-lifecycle-manager/packageserver-service LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.153\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1201 17:26:55.063314 6930 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.898013 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.908463 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.921649 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.934841 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.948031 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.961041 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.971590 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.971662 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.971675 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.971721 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.971737 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:55Z","lastTransitionTime":"2025-12-01T17:26:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.979342 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:55 crc kubenswrapper[4868]: I1201 17:26:55.993189 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cc7d784-77de-4109-ba8c-192bb74290f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c84b10381473d660a8159f66c6c2f012e0d46e588ef7f19a0f6950edc94806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://120b3cade14d3c00f0a22c5f4db0c5500a669bb07a267b65bf7a386938467f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://865f0cda78b8e8ef0cd3d52fd8c4baa2e0c0f67e9d2d89752ac7a7f5e5df0a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:55Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.074547 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.074592 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.074606 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.074623 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.074635 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:56Z","lastTransitionTime":"2025-12-01T17:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.174287 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.174397 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:56 crc kubenswrapper[4868]: E1201 17:26:56.174508 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:56 crc kubenswrapper[4868]: E1201 17:26:56.174661 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.178015 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.178056 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.178071 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.178091 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.178108 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:56Z","lastTransitionTime":"2025-12-01T17:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.280587 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.280664 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.280682 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.280711 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.280737 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:56Z","lastTransitionTime":"2025-12-01T17:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.383980 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.384031 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.384040 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.384059 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.384071 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:56Z","lastTransitionTime":"2025-12-01T17:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.487077 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.487148 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.487167 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.487194 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.487212 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:56Z","lastTransitionTime":"2025-12-01T17:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.670533 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.670577 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.670587 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.670604 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.670616 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:56Z","lastTransitionTime":"2025-12-01T17:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.746055 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovnkube-controller/3.log" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.772753 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.772806 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.772818 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.772837 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.772854 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:56Z","lastTransitionTime":"2025-12-01T17:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.875238 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.875274 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.875285 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.875301 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.875311 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:56Z","lastTransitionTime":"2025-12-01T17:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.977800 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.977844 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.977855 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.977873 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:56 crc kubenswrapper[4868]: I1201 17:26:56.977883 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:56Z","lastTransitionTime":"2025-12-01T17:26:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.080665 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.080710 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.080720 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.080740 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.080751 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:57Z","lastTransitionTime":"2025-12-01T17:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.171932 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.171995 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:57 crc kubenswrapper[4868]: E1201 17:26:57.172116 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:57 crc kubenswrapper[4868]: E1201 17:26:57.172268 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.182603 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.182662 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.182672 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.182689 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.182702 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:57Z","lastTransitionTime":"2025-12-01T17:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.286047 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.286101 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.286138 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.286160 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.286171 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:57Z","lastTransitionTime":"2025-12-01T17:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.390006 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.390067 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.390086 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.390103 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.390116 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:57Z","lastTransitionTime":"2025-12-01T17:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.492321 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.492367 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.492377 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.492394 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.492405 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:57Z","lastTransitionTime":"2025-12-01T17:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.600742 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.600794 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.600807 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.600826 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.600837 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:57Z","lastTransitionTime":"2025-12-01T17:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.702965 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.702997 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.703014 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.703033 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.703044 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:57Z","lastTransitionTime":"2025-12-01T17:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.805251 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.805300 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.805317 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.805334 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.805344 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:57Z","lastTransitionTime":"2025-12-01T17:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.907486 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.907527 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.907539 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.907559 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:57 crc kubenswrapper[4868]: I1201 17:26:57.907569 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:57Z","lastTransitionTime":"2025-12-01T17:26:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.010094 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.010141 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.010151 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.010170 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.010179 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:58Z","lastTransitionTime":"2025-12-01T17:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.112492 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.112531 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.112539 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.112554 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.112563 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:58Z","lastTransitionTime":"2025-12-01T17:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.171001 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.171189 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:26:58 crc kubenswrapper[4868]: E1201 17:26:58.171304 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:26:58 crc kubenswrapper[4868]: E1201 17:26:58.171500 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.185060 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.185061 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d8b1e3f6f1f6f8ed2844b1e533905130d255430f7031699ce02dfa6a3ac16a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:45Z\\\",\\\"message\\\":\\\"2025-12-01T17:26:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0\\\\n2025-12-01T17:26:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0 to /host/opt/cni/bin/\\\\n2025-12-01T17:26:00Z [verbose] multus-daemon started\\\\n2025-12-01T17:26:00Z [verbose] Readiness Indicator file check\\\\n2025-12-01T17:26:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.198133 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.212682 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.215052 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.215087 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.215097 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.215112 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.215123 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:58Z","lastTransitionTime":"2025-12-01T17:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.240548 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:24Z\\\",\\\"message\\\":\\\"io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00784edfb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:mhc-mtrc,Protocol:TCP,Port:8444,TargetPort:{1 0 mhc-mtrc},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: controller,},ClusterIP:10.217.4.167,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.167],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1201 17:26:24.043188 6530 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:55Z\\\",\\\"message\\\":\\\"t-ovn-kubernetes/ovnkube-node-q9z6l\\\\nI1201 17:26:55.063250 6930 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1201 17:26:55.063211 6930 services_controller.go:356] Processing sync for service default/kubernetes for network=default\\\\nI1201 17:26:55.063161 6930 services_controller.go:443] Built service openshift-operator-lifecycle-manager/packageserver-service LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.153\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1201 17:26:55.063314 6930 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.256129 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.270099 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.286142 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.304403 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.319438 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.319497 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.319515 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.319543 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.319498 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.319563 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:58Z","lastTransitionTime":"2025-12-01T17:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.336590 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.357181 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.370916 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cc7d784-77de-4109-ba8c-192bb74290f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c84b10381473d660a8159f66c6c2f012e0d46e588ef7f19a0f6950edc94806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://120b3cade14d3c00f0a22c5f4db0c5500a669bb07a267b65bf7a386938467f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://865f0cda78b8e8ef0cd3d52fd8c4baa2e0c0f67e9d2d89752ac7a7f5e5df0a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.387409 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.421561 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.423120 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.423179 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.423190 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.423209 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.423220 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:58Z","lastTransitionTime":"2025-12-01T17:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.450445 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.473266 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.488860 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:26:58Z is after 2025-08-24T17:21:41Z" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.526058 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.526103 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.526113 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.526129 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.526140 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:58Z","lastTransitionTime":"2025-12-01T17:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.628189 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.628237 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.628248 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.628265 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.628276 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:58Z","lastTransitionTime":"2025-12-01T17:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.731529 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.731601 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.731615 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.731635 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.731653 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:58Z","lastTransitionTime":"2025-12-01T17:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.834813 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.834871 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.834887 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.834910 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.834927 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:58Z","lastTransitionTime":"2025-12-01T17:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.938028 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.938098 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.938198 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.938231 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:58 crc kubenswrapper[4868]: I1201 17:26:58.938247 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:58Z","lastTransitionTime":"2025-12-01T17:26:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.040742 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.040789 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.040800 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.040819 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.040832 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:59Z","lastTransitionTime":"2025-12-01T17:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.143205 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.143245 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.143259 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.143275 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.143285 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:59Z","lastTransitionTime":"2025-12-01T17:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.171518 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:26:59 crc kubenswrapper[4868]: E1201 17:26:59.171667 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.171543 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:26:59 crc kubenswrapper[4868]: E1201 17:26:59.172464 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.246036 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.246086 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.246095 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.246113 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.246123 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:59Z","lastTransitionTime":"2025-12-01T17:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.349384 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.349445 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.349458 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.349483 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.349494 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:59Z","lastTransitionTime":"2025-12-01T17:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.452794 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.452871 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.452890 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.452918 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.452977 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:59Z","lastTransitionTime":"2025-12-01T17:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.555706 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.555744 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.555754 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.555770 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.555779 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:59Z","lastTransitionTime":"2025-12-01T17:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.658834 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.658912 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.658932 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.658986 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.659009 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:59Z","lastTransitionTime":"2025-12-01T17:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.761181 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.761252 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.761271 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.761299 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.761319 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:59Z","lastTransitionTime":"2025-12-01T17:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.864587 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.864637 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.864649 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.864674 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.864688 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:59Z","lastTransitionTime":"2025-12-01T17:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.968838 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.968904 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.968916 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.968934 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:26:59 crc kubenswrapper[4868]: I1201 17:26:59.968975 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:26:59Z","lastTransitionTime":"2025-12-01T17:26:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.072110 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.072199 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.072221 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.072254 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.072275 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:00Z","lastTransitionTime":"2025-12-01T17:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.171364 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.171501 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:00 crc kubenswrapper[4868]: E1201 17:27:00.171594 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:00 crc kubenswrapper[4868]: E1201 17:27:00.171719 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.175453 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.175530 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.175555 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.175588 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.175613 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:00Z","lastTransitionTime":"2025-12-01T17:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.278555 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.278992 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.279007 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.279027 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.279037 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:00Z","lastTransitionTime":"2025-12-01T17:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.381844 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.381887 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.381900 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.381921 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.381934 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:00Z","lastTransitionTime":"2025-12-01T17:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.484907 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.484981 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.484997 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.485019 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.485035 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:00Z","lastTransitionTime":"2025-12-01T17:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.587558 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.587602 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.587611 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.587628 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.587638 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:00Z","lastTransitionTime":"2025-12-01T17:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.690486 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.690602 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.690640 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.690678 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.690700 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:00Z","lastTransitionTime":"2025-12-01T17:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.795787 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.796154 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.796181 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.796249 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.796274 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:00Z","lastTransitionTime":"2025-12-01T17:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.899569 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.899638 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.899654 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.899676 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:00 crc kubenswrapper[4868]: I1201 17:27:00.899692 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:00Z","lastTransitionTime":"2025-12-01T17:27:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.003505 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.003592 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.003612 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.003644 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.003678 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:01Z","lastTransitionTime":"2025-12-01T17:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.106620 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.106672 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.106685 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.106706 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.106719 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:01Z","lastTransitionTime":"2025-12-01T17:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.171058 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.171088 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:01 crc kubenswrapper[4868]: E1201 17:27:01.171346 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:01 crc kubenswrapper[4868]: E1201 17:27:01.171450 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.209609 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.209663 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.209675 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.209692 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.209704 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:01Z","lastTransitionTime":"2025-12-01T17:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.312633 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.312691 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.312702 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.312730 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.312745 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:01Z","lastTransitionTime":"2025-12-01T17:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.415054 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.415101 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.415109 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.415123 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.415133 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:01Z","lastTransitionTime":"2025-12-01T17:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.518051 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.518097 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.518111 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.518128 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.518137 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:01Z","lastTransitionTime":"2025-12-01T17:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.620668 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.620726 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.620743 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.620771 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.620787 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:01Z","lastTransitionTime":"2025-12-01T17:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.724648 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.724686 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.724695 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.724712 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.724720 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:01Z","lastTransitionTime":"2025-12-01T17:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.828383 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.828466 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.828485 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.828514 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.828534 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:01Z","lastTransitionTime":"2025-12-01T17:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.931930 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.932037 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.932058 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.932085 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:01 crc kubenswrapper[4868]: I1201 17:27:01.932103 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:01Z","lastTransitionTime":"2025-12-01T17:27:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.035527 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.035598 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.035616 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.035642 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.035665 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:02Z","lastTransitionTime":"2025-12-01T17:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.138436 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.138545 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.138569 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.138601 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.138628 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:02Z","lastTransitionTime":"2025-12-01T17:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.171722 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.171744 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:02 crc kubenswrapper[4868]: E1201 17:27:02.172485 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:02 crc kubenswrapper[4868]: E1201 17:27:02.172709 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.241643 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.241696 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.241706 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.241723 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.241734 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:02Z","lastTransitionTime":"2025-12-01T17:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.344368 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.344423 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.344434 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.344454 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.344468 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:02Z","lastTransitionTime":"2025-12-01T17:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.448710 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.448776 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.448795 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.448823 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.448844 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:02Z","lastTransitionTime":"2025-12-01T17:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.551780 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.551835 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.551849 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.551869 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.551886 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:02Z","lastTransitionTime":"2025-12-01T17:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.664836 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.664899 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.664912 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.664968 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.664985 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:02Z","lastTransitionTime":"2025-12-01T17:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.768250 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.768288 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.768298 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.768315 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.768327 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:02Z","lastTransitionTime":"2025-12-01T17:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.871252 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.871296 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.871307 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.871323 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.871332 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:02Z","lastTransitionTime":"2025-12-01T17:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.973630 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.973670 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.973678 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.973695 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:02 crc kubenswrapper[4868]: I1201 17:27:02.973705 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:02Z","lastTransitionTime":"2025-12-01T17:27:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.076175 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.076469 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.076571 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.076661 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.076747 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:03Z","lastTransitionTime":"2025-12-01T17:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.171639 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.171693 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:03 crc kubenswrapper[4868]: E1201 17:27:03.171773 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:03 crc kubenswrapper[4868]: E1201 17:27:03.171837 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.179408 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.179441 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.179451 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.179466 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.179476 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:03Z","lastTransitionTime":"2025-12-01T17:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.281661 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.281707 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.281718 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.281737 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.281749 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:03Z","lastTransitionTime":"2025-12-01T17:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.384536 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.385146 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.385238 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.385336 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.385405 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:03Z","lastTransitionTime":"2025-12-01T17:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.488043 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.488338 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.488593 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.488795 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.489009 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:03Z","lastTransitionTime":"2025-12-01T17:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.591648 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.591686 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.591697 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.591733 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.591743 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:03Z","lastTransitionTime":"2025-12-01T17:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.611155 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.611198 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.611213 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.611231 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.611242 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:03Z","lastTransitionTime":"2025-12-01T17:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:03 crc kubenswrapper[4868]: E1201 17:27:03.624083 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.629118 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.629336 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.629407 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.629484 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.629552 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:03Z","lastTransitionTime":"2025-12-01T17:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:03 crc kubenswrapper[4868]: E1201 17:27:03.643495 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.647193 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.647673 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.647749 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.647835 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.647917 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:03Z","lastTransitionTime":"2025-12-01T17:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:03 crc kubenswrapper[4868]: E1201 17:27:03.668331 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.672811 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.672851 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.672863 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.672880 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.672892 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:03Z","lastTransitionTime":"2025-12-01T17:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:03 crc kubenswrapper[4868]: E1201 17:27:03.686271 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.690046 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.690089 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.690102 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.690127 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.690140 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:03Z","lastTransitionTime":"2025-12-01T17:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:03 crc kubenswrapper[4868]: E1201 17:27:03.703145 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:03Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:03 crc kubenswrapper[4868]: E1201 17:27:03.703268 4868 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.705160 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.705197 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.705211 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.705232 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.705246 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:03Z","lastTransitionTime":"2025-12-01T17:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.808690 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.808770 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.808795 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.808830 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.808853 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:03Z","lastTransitionTime":"2025-12-01T17:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.911860 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.912723 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.912915 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.913100 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:03 crc kubenswrapper[4868]: I1201 17:27:03.913228 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:03Z","lastTransitionTime":"2025-12-01T17:27:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.016374 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.016450 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.016465 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.016500 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.016519 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:04Z","lastTransitionTime":"2025-12-01T17:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.118855 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.118918 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.118930 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.118971 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.118991 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:04Z","lastTransitionTime":"2025-12-01T17:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.171179 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.171252 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:04 crc kubenswrapper[4868]: E1201 17:27:04.171405 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:04 crc kubenswrapper[4868]: E1201 17:27:04.171528 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.221183 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.221257 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.221270 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.221290 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.221303 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:04Z","lastTransitionTime":"2025-12-01T17:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.323706 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.323761 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.323773 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.323796 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.323809 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:04Z","lastTransitionTime":"2025-12-01T17:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.426037 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.426084 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.426100 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.426123 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.426138 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:04Z","lastTransitionTime":"2025-12-01T17:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.528596 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.528675 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.528695 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.528718 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.528735 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:04Z","lastTransitionTime":"2025-12-01T17:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.630864 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.630931 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.630973 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.631002 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.631022 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:04Z","lastTransitionTime":"2025-12-01T17:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.734823 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.734906 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.734921 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.734987 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.735009 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:04Z","lastTransitionTime":"2025-12-01T17:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.837410 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.837488 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.837506 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.837534 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.837555 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:04Z","lastTransitionTime":"2025-12-01T17:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.940878 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.940994 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.941014 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.941042 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:04 crc kubenswrapper[4868]: I1201 17:27:04.941062 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:04Z","lastTransitionTime":"2025-12-01T17:27:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.043207 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.043563 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.043687 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.043839 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.043992 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:05Z","lastTransitionTime":"2025-12-01T17:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.147329 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.147433 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.147462 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.147496 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.147523 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:05Z","lastTransitionTime":"2025-12-01T17:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.171840 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:05 crc kubenswrapper[4868]: E1201 17:27:05.172017 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.172143 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:05 crc kubenswrapper[4868]: E1201 17:27:05.172383 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.251898 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.251971 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.251984 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.252005 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.252019 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:05Z","lastTransitionTime":"2025-12-01T17:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.355991 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.356063 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.356087 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.356115 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.356135 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:05Z","lastTransitionTime":"2025-12-01T17:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.458328 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.458383 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.458398 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.458420 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.458433 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:05Z","lastTransitionTime":"2025-12-01T17:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.561039 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.561116 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.561140 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.561173 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.561200 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:05Z","lastTransitionTime":"2025-12-01T17:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.664037 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.664077 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.664087 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.664112 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.664123 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:05Z","lastTransitionTime":"2025-12-01T17:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.766686 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.766730 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.766739 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.766754 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.766765 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:05Z","lastTransitionTime":"2025-12-01T17:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.868692 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.868788 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.868814 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.868844 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.868865 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:05Z","lastTransitionTime":"2025-12-01T17:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.971917 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.972059 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.972081 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.972111 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:05 crc kubenswrapper[4868]: I1201 17:27:05.972129 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:05Z","lastTransitionTime":"2025-12-01T17:27:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.079583 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.079673 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.079701 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.079740 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.079774 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:06Z","lastTransitionTime":"2025-12-01T17:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.171814 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.171965 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:06 crc kubenswrapper[4868]: E1201 17:27:06.172122 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:06 crc kubenswrapper[4868]: E1201 17:27:06.172291 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.182727 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.182790 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.182808 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.182829 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.182844 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:06Z","lastTransitionTime":"2025-12-01T17:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.286770 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.286854 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.286875 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.286905 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.286924 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:06Z","lastTransitionTime":"2025-12-01T17:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.391145 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.391271 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.391299 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.391333 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.391355 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:06Z","lastTransitionTime":"2025-12-01T17:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.494979 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.495047 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.495069 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.495098 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.495119 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:06Z","lastTransitionTime":"2025-12-01T17:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.599228 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.599337 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.599362 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.599391 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.599411 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:06Z","lastTransitionTime":"2025-12-01T17:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.702450 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.702492 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.702501 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.702517 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.702528 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:06Z","lastTransitionTime":"2025-12-01T17:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.805587 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.805737 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.805763 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.805785 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.805801 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:06Z","lastTransitionTime":"2025-12-01T17:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.909254 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.909328 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.909365 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.909402 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:06 crc kubenswrapper[4868]: I1201 17:27:06.909427 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:06Z","lastTransitionTime":"2025-12-01T17:27:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.013603 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.013678 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.013702 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.013734 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.013756 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:07Z","lastTransitionTime":"2025-12-01T17:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.117696 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.117778 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.117799 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.117827 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.117847 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:07Z","lastTransitionTime":"2025-12-01T17:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.171585 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.171585 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:07 crc kubenswrapper[4868]: E1201 17:27:07.171849 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:07 crc kubenswrapper[4868]: E1201 17:27:07.172059 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.221891 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.221982 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.222004 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.222029 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.222046 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:07Z","lastTransitionTime":"2025-12-01T17:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.325129 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.325215 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.325239 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.325274 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.325302 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:07Z","lastTransitionTime":"2025-12-01T17:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.428983 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.429051 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.429067 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.429091 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.429109 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:07Z","lastTransitionTime":"2025-12-01T17:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.532219 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.532274 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.532285 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.532304 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.532323 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:07Z","lastTransitionTime":"2025-12-01T17:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.636886 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.636986 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.636998 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.637023 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.637038 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:07Z","lastTransitionTime":"2025-12-01T17:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.741420 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.741508 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.741539 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.741576 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.741604 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:07Z","lastTransitionTime":"2025-12-01T17:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.845305 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.845387 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.845410 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.845442 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.845465 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:07Z","lastTransitionTime":"2025-12-01T17:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.949295 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.949347 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.949363 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.949390 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:07 crc kubenswrapper[4868]: I1201 17:27:07.949408 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:07Z","lastTransitionTime":"2025-12-01T17:27:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.052847 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.052919 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.052996 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.053036 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.053063 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:08Z","lastTransitionTime":"2025-12-01T17:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.157564 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.157746 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.157822 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.157858 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.157929 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:08Z","lastTransitionTime":"2025-12-01T17:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.170925 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.171026 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:08 crc kubenswrapper[4868]: E1201 17:27:08.171180 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:08 crc kubenswrapper[4868]: E1201 17:27:08.171512 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.173214 4868 scope.go:117] "RemoveContainer" containerID="fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c" Dec 01 17:27:08 crc kubenswrapper[4868]: E1201 17:27:08.173715 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-q9z6l_openshift-ovn-kubernetes(5baca705-90cc-4198-b641-214a2b915086)\"" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podUID="5baca705-90cc-4198-b641-214a2b915086" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.193790 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cc7d784-77de-4109-ba8c-192bb74290f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c84b10381473d660a8159f66c6c2f012e0d46e588ef7f19a0f6950edc94806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://120b3cade14d3c00f0a22c5f4db0c5500a669bb07a267b65bf7a386938467f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://865f0cda78b8e8ef0cd3d52fd8c4baa2e0c0f67e9d2d89752ac7a7f5e5df0a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.215097 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.236193 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.255330 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.261906 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.262161 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.262209 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.262270 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.262298 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:08Z","lastTransitionTime":"2025-12-01T17:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.274598 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.299485 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2b3e938-c83a-4a6d-aa5e-3f6c460f8468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1280eab687f307ac45cc184525a9aa1737ef93d7b0b1921dcb474c0d203f7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cca7f9573182f19dae9f87074d72f5ea36842aea72907153a7789b33eaee86b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cca7f9573182f19dae9f87074d72f5ea36842aea72907153a7789b33eaee86b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.316679 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.332167 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.351924 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.365968 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.366043 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.366060 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.366109 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.366155 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:08Z","lastTransitionTime":"2025-12-01T17:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.370469 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d8b1e3f6f1f6f8ed2844b1e533905130d255430f7031699ce02dfa6a3ac16a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:45Z\\\",\\\"message\\\":\\\"2025-12-01T17:26:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0\\\\n2025-12-01T17:26:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0 to /host/opt/cni/bin/\\\\n2025-12-01T17:26:00Z [verbose] multus-daemon started\\\\n2025-12-01T17:26:00Z [verbose] Readiness Indicator file check\\\\n2025-12-01T17:26:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.386325 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.402266 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.416512 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.433112 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.449366 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.463091 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.475175 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.475274 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.475297 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.475357 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.475379 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:08Z","lastTransitionTime":"2025-12-01T17:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.484839 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.511154 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1892a55c56eff6857b6f51ee4f70a285c59f3b0ee860e12d709eedc6ccc7cbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:24Z\\\",\\\"message\\\":\\\"io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00784edfb \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:mhc-mtrc,Protocol:TCP,Port:8444,TargetPort:{1 0 mhc-mtrc},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: controller,},ClusterIP:10.217.4.167,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.167],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1201 17:26:24.043188 6530 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:55Z\\\",\\\"message\\\":\\\"t-ovn-kubernetes/ovnkube-node-q9z6l\\\\nI1201 17:26:55.063250 6930 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1201 17:26:55.063211 6930 services_controller.go:356] Processing sync for service default/kubernetes for network=default\\\\nI1201 17:26:55.063161 6930 services_controller.go:443] Built service openshift-operator-lifecycle-manager/packageserver-service LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.153\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1201 17:26:55.063314 6930 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.529466 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.546434 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7e1f752-fd18-475d-af3b-8af3a51bc2d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4e129aacdeaaa6ff3daf6a35fab235e899d2a41f5007d752b4741ff8e2a62df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1421765ce4e20677884c56cea8013947804ae3734825a409c6afa2426aa2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdnmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c7p72\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.561219 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2b3e938-c83a-4a6d-aa5e-3f6c460f8468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1280eab687f307ac45cc184525a9aa1737ef93d7b0b1921dcb474c0d203f7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cca7f9573182f19dae9f87074d72f5ea36842aea72907153a7789b33eaee86b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cca7f9573182f19dae9f87074d72f5ea36842aea72907153a7789b33eaee86b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.574270 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T17:25:46Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 17:25:40.900286 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 17:25:40.903490 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2089004685/tls.crt::/tmp/serving-cert-2089004685/tls.key\\\\\\\"\\\\nI1201 17:25:46.520039 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 17:25:46.523528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 17:25:46.523555 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 17:25:46.523585 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 17:25:46.523592 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 17:25:46.533800 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 17:25:46.533897 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533922 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 17:25:46.533965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 17:25:46.533995 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 17:25:46.534015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 17:25:46.534034 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 17:25:46.533830 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 17:25:46.536239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.583378 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.583415 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.583427 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.583448 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.583461 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:08Z","lastTransitionTime":"2025-12-01T17:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.587406 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4e08bf-df56-4490-8e79-a4175c727405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e51f818424d3336cf888fadac37b5b414f20d108e6f4acb82c1d9ca0640b8c2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78dkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-njgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.605568 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9tlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ce688b4-36b1-48d4-a6d1-8aec723125c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d8b1e3f6f1f6f8ed2844b1e533905130d255430f7031699ce02dfa6a3ac16a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:45Z\\\",\\\"message\\\":\\\"2025-12-01T17:26:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0\\\\n2025-12-01T17:26:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6551cb84-faa3-4910-8046-df25e17963f0 to /host/opt/cni/bin/\\\\n2025-12-01T17:26:00Z [verbose] multus-daemon started\\\\n2025-12-01T17:26:00Z [verbose] Readiness Indicator file check\\\\n2025-12-01T17:26:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9mv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9tlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.618966 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c7mwj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76081c1f-3c70-497b-8591-53350dc17405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://867e6d077ee9b53aaafb7eeb216f097f8acb01a3964fc2073dba987df6ca8681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-729qk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c7mwj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.638305 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hw725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46acee96-a86d-4abf-a79f-7cfc61fbab92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5a79321f718cb0c6b938983c8c096c44cdab5b53e04a9c97f516fa917b0a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d65b24e799adf489955ad70b58fb80394cee321bd44a545caaadf225b199fbfd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a19a80298c539fd7901330829e97bd10a2601c861bc9d3667cb9b8a12883246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://144a45525fbfbbf23023edc529454e44439635e697878dceb8a7805578376c21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://176d923a5c5065f5e7ff257c26d6d347e8fd1b7fd53baf3cec24006b715a9bab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://385a648160eab0893f766dd6e801f21166a390ce453066498cd96a167e7580ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9931124adb907822774ce5f8d31c8284976c58b3818e7f156a75215ad7a2d886\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:26:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qk4sq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hw725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.659587 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5baca705-90cc-4198-b641-214a2b915086\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T17:26:55Z\\\",\\\"message\\\":\\\"t-ovn-kubernetes/ovnkube-node-q9z6l\\\\nI1201 17:26:55.063250 6930 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1201 17:26:55.063211 6930 services_controller.go:356] Processing sync for service default/kubernetes for network=default\\\\nI1201 17:26:55.063161 6930 services_controller.go:443] Built service openshift-operator-lifecycle-manager/packageserver-service LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.153\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1201 17:26:55.063314 6930 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T17:26:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-q9z6l_openshift-ovn-kubernetes(5baca705-90cc-4198-b641-214a2b915086)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j2cpz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-q9z6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.672670 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-64mjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b184c651-661a-4d84-bb0a-75107bf5de1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c46110bb23defcd5e956717727ceaf44c1a2f107302dce81e6c80ee318edcd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:26:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t68cm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-64mjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.684910 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4afd9a0c-e8c7-42ec-9771-036c3441cbce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zc4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:26:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nxfkx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.685894 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.685926 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.685937 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.685975 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.685985 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:08Z","lastTransitionTime":"2025-12-01T17:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.700428 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e5f25a4-fa05-49ee-80f2-667adcce1a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542189a3fd90281791a96757928a22471a6f13d625679464fe69439dccbc5804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0832598f00a75da074c93048f29d54cb48c45d259a9e79a1b9cb4c7076da2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aca0ff93fb53caae10c758c9fb0930854f72ba0dfccac392c3941add0c6174b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.718355 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41f4e59ffcd4c7304929018c5a0642427c2c90d139ea6691338f2e3bdf8a9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://760748214ae5f334096b9dd0399590e9dbb04cdb52b6c820c87480b7a1576b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.737072 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c22f68483f1d9222a36a91c74bb9e570cbece0a9eee93361255ae1feb528c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.752061 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd73c46ba92066fba504fff85e7d71ea5be999a57680469313d2e338d81ab25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.769852 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.789159 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cc7d784-77de-4109-ba8c-192bb74290f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:26:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c84b10381473d660a8159f66c6c2f012e0d46e588ef7f19a0f6950edc94806\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://120b3cade14d3c00f0a22c5f4db0c5500a669bb07a267b65bf7a386938467f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://865f0cda78b8e8ef0cd3d52fd8c4baa2e0c0f67e9d2d89752ac7a7f5e5df0a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bb715396052af0776ece1f31b7492f739731513f4a7a49cb1a8d9b939bd066\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:25:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:25:29Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T17:25:28Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.789348 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.789397 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.789409 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.789429 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.789444 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:08Z","lastTransitionTime":"2025-12-01T17:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.802872 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:25:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:08Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.892784 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.892842 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.892855 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.892875 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.892887 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:08Z","lastTransitionTime":"2025-12-01T17:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.995730 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.995776 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.995788 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.995808 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:08 crc kubenswrapper[4868]: I1201 17:27:08.995819 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:08Z","lastTransitionTime":"2025-12-01T17:27:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.100250 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.100315 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.100334 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.100356 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.100371 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:09Z","lastTransitionTime":"2025-12-01T17:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.171705 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.171864 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:09 crc kubenswrapper[4868]: E1201 17:27:09.171999 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:09 crc kubenswrapper[4868]: E1201 17:27:09.172932 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.204601 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.205147 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.205189 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.205218 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.205236 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:09Z","lastTransitionTime":"2025-12-01T17:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.308553 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.308590 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.308599 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.308616 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.308627 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:09Z","lastTransitionTime":"2025-12-01T17:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.411767 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.411838 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.411857 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.411883 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.411907 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:09Z","lastTransitionTime":"2025-12-01T17:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.515114 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.515192 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.515208 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.515232 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.515253 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:09Z","lastTransitionTime":"2025-12-01T17:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.618256 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.618325 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.618350 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.618383 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.618408 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:09Z","lastTransitionTime":"2025-12-01T17:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.721980 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.722028 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.722040 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.722059 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.722072 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:09Z","lastTransitionTime":"2025-12-01T17:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.825614 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.825666 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.825691 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.825721 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.825739 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:09Z","lastTransitionTime":"2025-12-01T17:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.929536 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.930048 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.930208 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.930414 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:09 crc kubenswrapper[4868]: I1201 17:27:09.930572 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:09Z","lastTransitionTime":"2025-12-01T17:27:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.033744 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.033785 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.033794 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.033810 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.033825 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:10Z","lastTransitionTime":"2025-12-01T17:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.137693 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.138166 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.138365 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.138605 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.138765 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:10Z","lastTransitionTime":"2025-12-01T17:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.171841 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.171928 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:10 crc kubenswrapper[4868]: E1201 17:27:10.172079 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:10 crc kubenswrapper[4868]: E1201 17:27:10.172240 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.242970 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.243066 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.243082 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.243106 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.243125 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:10Z","lastTransitionTime":"2025-12-01T17:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.346295 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.346366 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.346380 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.346401 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.346417 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:10Z","lastTransitionTime":"2025-12-01T17:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.450735 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.450824 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.450838 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.450856 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.450892 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:10Z","lastTransitionTime":"2025-12-01T17:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.554446 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.554542 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.554573 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.554610 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.554641 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:10Z","lastTransitionTime":"2025-12-01T17:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.658551 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.658648 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.658677 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.658712 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.658737 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:10Z","lastTransitionTime":"2025-12-01T17:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.762100 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.762157 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.762230 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.762253 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.762265 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:10Z","lastTransitionTime":"2025-12-01T17:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.865004 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.865062 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.865078 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.865098 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.865112 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:10Z","lastTransitionTime":"2025-12-01T17:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.968239 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.968288 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.968298 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.968318 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:10 crc kubenswrapper[4868]: I1201 17:27:10.968329 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:10Z","lastTransitionTime":"2025-12-01T17:27:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.070893 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.071011 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.071033 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.071062 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.071083 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:11Z","lastTransitionTime":"2025-12-01T17:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.171260 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.171260 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:11 crc kubenswrapper[4868]: E1201 17:27:11.171615 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:11 crc kubenswrapper[4868]: E1201 17:27:11.172354 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.173614 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.173666 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.173679 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.173699 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.173713 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:11Z","lastTransitionTime":"2025-12-01T17:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.277640 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.277729 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.277741 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.277785 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.277800 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:11Z","lastTransitionTime":"2025-12-01T17:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.381193 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.381274 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.381300 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.381328 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.381345 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:11Z","lastTransitionTime":"2025-12-01T17:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.484213 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.484311 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.484338 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.484369 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.484391 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:11Z","lastTransitionTime":"2025-12-01T17:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.586834 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.586918 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.586959 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.586980 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.586991 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:11Z","lastTransitionTime":"2025-12-01T17:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.690025 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.690079 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.690097 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.690122 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.690141 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:11Z","lastTransitionTime":"2025-12-01T17:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.793776 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.793870 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.793889 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.793919 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.793997 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:11Z","lastTransitionTime":"2025-12-01T17:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.896901 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.896971 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.896981 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.896995 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:11 crc kubenswrapper[4868]: I1201 17:27:11.897004 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:11Z","lastTransitionTime":"2025-12-01T17:27:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.000322 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.000363 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.000371 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.000385 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.000394 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:12Z","lastTransitionTime":"2025-12-01T17:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.104412 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.104494 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.104512 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.104540 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.104562 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:12Z","lastTransitionTime":"2025-12-01T17:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.171329 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.171435 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:12 crc kubenswrapper[4868]: E1201 17:27:12.171611 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:12 crc kubenswrapper[4868]: E1201 17:27:12.171908 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.194857 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.208296 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.208326 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.208338 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.208353 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.208366 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:12Z","lastTransitionTime":"2025-12-01T17:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.311711 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.311756 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.311767 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.311787 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.311799 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:12Z","lastTransitionTime":"2025-12-01T17:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.414912 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.415000 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.415018 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.415046 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.415063 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:12Z","lastTransitionTime":"2025-12-01T17:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.519928 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.520017 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.520031 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.520053 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.520065 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:12Z","lastTransitionTime":"2025-12-01T17:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.624346 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.624413 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.624428 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.624449 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.624467 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:12Z","lastTransitionTime":"2025-12-01T17:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.728171 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.728222 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.728236 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.728257 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.728272 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:12Z","lastTransitionTime":"2025-12-01T17:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.744567 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs\") pod \"network-metrics-daemon-nxfkx\" (UID: \"4afd9a0c-e8c7-42ec-9771-036c3441cbce\") " pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:12 crc kubenswrapper[4868]: E1201 17:27:12.744819 4868 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:27:12 crc kubenswrapper[4868]: E1201 17:27:12.744987 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs podName:4afd9a0c-e8c7-42ec-9771-036c3441cbce nodeName:}" failed. No retries permitted until 2025-12-01 17:28:16.744924274 +0000 UTC m=+169.116034685 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs") pod "network-metrics-daemon-nxfkx" (UID: "4afd9a0c-e8c7-42ec-9771-036c3441cbce") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.831678 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.831730 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.831740 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.831757 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.831767 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:12Z","lastTransitionTime":"2025-12-01T17:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.934780 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.934865 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.934883 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.934909 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:12 crc kubenswrapper[4868]: I1201 17:27:12.934928 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:12Z","lastTransitionTime":"2025-12-01T17:27:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.037845 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.037903 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.037919 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.037953 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.037967 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:13Z","lastTransitionTime":"2025-12-01T17:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.141421 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.141527 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.141550 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.141575 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.141597 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:13Z","lastTransitionTime":"2025-12-01T17:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.171443 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:13 crc kubenswrapper[4868]: E1201 17:27:13.171644 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.171458 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:13 crc kubenswrapper[4868]: E1201 17:27:13.171850 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.244421 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.244484 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.244499 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.244521 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.244536 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:13Z","lastTransitionTime":"2025-12-01T17:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.348192 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.348276 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.348288 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.348305 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.348342 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:13Z","lastTransitionTime":"2025-12-01T17:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.451542 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.451612 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.451633 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.451661 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.451681 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:13Z","lastTransitionTime":"2025-12-01T17:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.555109 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.555226 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.555249 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.555276 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.555297 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:13Z","lastTransitionTime":"2025-12-01T17:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.658410 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.658453 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.658464 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.658478 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.658489 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:13Z","lastTransitionTime":"2025-12-01T17:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.761743 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.761844 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.761890 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.761922 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.761981 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:13Z","lastTransitionTime":"2025-12-01T17:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.865341 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.865446 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.865468 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.865501 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.865526 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:13Z","lastTransitionTime":"2025-12-01T17:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.888332 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.888387 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.888404 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.888422 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.888439 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:13Z","lastTransitionTime":"2025-12-01T17:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:13 crc kubenswrapper[4868]: E1201 17:27:13.901871 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:13Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.907640 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.907693 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.907705 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.907724 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.907739 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:13Z","lastTransitionTime":"2025-12-01T17:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:13 crc kubenswrapper[4868]: E1201 17:27:13.924063 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:13Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.930019 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.930090 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.930109 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.930132 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.930147 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:13Z","lastTransitionTime":"2025-12-01T17:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:13 crc kubenswrapper[4868]: E1201 17:27:13.944160 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:13Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.955899 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.955973 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.955985 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.956061 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.956114 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:13Z","lastTransitionTime":"2025-12-01T17:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:13 crc kubenswrapper[4868]: E1201 17:27:13.973293 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:13Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.978628 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.978695 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.978709 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.978731 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:13 crc kubenswrapper[4868]: I1201 17:27:13.978743 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:13Z","lastTransitionTime":"2025-12-01T17:27:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:13 crc kubenswrapper[4868]: E1201 17:27:13.997881 4868 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T17:27:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"03b0a5d8-65dc-4356-bde6-5907df7dc763\\\",\\\"systemUUID\\\":\\\"ba6f0959-f434-42df-bfe4-56be307a3fb1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T17:27:13Z is after 2025-08-24T17:21:41Z" Dec 01 17:27:13 crc kubenswrapper[4868]: E1201 17:27:13.998025 4868 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.000006 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.000031 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.000040 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.000057 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.000071 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:14Z","lastTransitionTime":"2025-12-01T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.105106 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.105185 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.105216 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.105253 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.105274 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:14Z","lastTransitionTime":"2025-12-01T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.171843 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.171917 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:14 crc kubenswrapper[4868]: E1201 17:27:14.172080 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:14 crc kubenswrapper[4868]: E1201 17:27:14.172238 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.208598 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.208650 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.208664 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.208688 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.208704 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:14Z","lastTransitionTime":"2025-12-01T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.311731 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.311785 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.311795 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.311813 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.311824 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:14Z","lastTransitionTime":"2025-12-01T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.414665 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.414734 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.414751 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.414778 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.414794 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:14Z","lastTransitionTime":"2025-12-01T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.518509 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.518586 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.518612 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.518645 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.518670 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:14Z","lastTransitionTime":"2025-12-01T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.624213 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.624314 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.624333 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.624362 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.624380 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:14Z","lastTransitionTime":"2025-12-01T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.727142 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.727209 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.727233 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.727265 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.727291 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:14Z","lastTransitionTime":"2025-12-01T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.830263 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.830424 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.830446 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.830473 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.830495 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:14Z","lastTransitionTime":"2025-12-01T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.933406 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.933471 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.933500 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.933528 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:14 crc kubenswrapper[4868]: I1201 17:27:14.933556 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:14Z","lastTransitionTime":"2025-12-01T17:27:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.037076 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.037147 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.037169 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.037205 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.037229 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:15Z","lastTransitionTime":"2025-12-01T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.140405 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.140469 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.140485 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.140508 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.140524 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:15Z","lastTransitionTime":"2025-12-01T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.171505 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.171532 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:15 crc kubenswrapper[4868]: E1201 17:27:15.171760 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:15 crc kubenswrapper[4868]: E1201 17:27:15.171881 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.242982 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.243037 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.243048 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.243064 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.243075 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:15Z","lastTransitionTime":"2025-12-01T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.346407 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.346464 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.346473 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.346492 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.346512 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:15Z","lastTransitionTime":"2025-12-01T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.451283 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.451341 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.451357 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.451380 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.451419 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:15Z","lastTransitionTime":"2025-12-01T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.554983 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.555039 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.555049 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.555067 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.555079 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:15Z","lastTransitionTime":"2025-12-01T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.658283 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.658337 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.658365 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.658393 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.658411 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:15Z","lastTransitionTime":"2025-12-01T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.761768 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.761833 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.761843 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.761862 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.761876 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:15Z","lastTransitionTime":"2025-12-01T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.864480 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.864529 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.864537 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.864551 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.864565 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:15Z","lastTransitionTime":"2025-12-01T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.967612 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.967656 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.967667 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.967682 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:15 crc kubenswrapper[4868]: I1201 17:27:15.967695 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:15Z","lastTransitionTime":"2025-12-01T17:27:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.071301 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.071362 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.071381 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.071401 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.071413 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:16Z","lastTransitionTime":"2025-12-01T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.171909 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.172066 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:16 crc kubenswrapper[4868]: E1201 17:27:16.172513 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:16 crc kubenswrapper[4868]: E1201 17:27:16.172598 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.174187 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.174234 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.174249 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.174270 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.174287 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:16Z","lastTransitionTime":"2025-12-01T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.277673 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.277741 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.277755 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.277773 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.277786 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:16Z","lastTransitionTime":"2025-12-01T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.387729 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.388270 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.388365 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.388404 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.388431 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:16Z","lastTransitionTime":"2025-12-01T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.491345 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.491392 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.491408 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.491434 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.491451 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:16Z","lastTransitionTime":"2025-12-01T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.594698 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.594771 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.594789 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.594815 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.594834 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:16Z","lastTransitionTime":"2025-12-01T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.697688 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.697743 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.697756 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.697777 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.697791 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:16Z","lastTransitionTime":"2025-12-01T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.801367 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.801440 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.801464 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.801495 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.801521 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:16Z","lastTransitionTime":"2025-12-01T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.905016 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.905066 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.905083 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.905104 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:16 crc kubenswrapper[4868]: I1201 17:27:16.905119 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:16Z","lastTransitionTime":"2025-12-01T17:27:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.008022 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.008097 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.008122 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.008152 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.008171 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:17Z","lastTransitionTime":"2025-12-01T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.111313 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.111389 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.111412 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.111443 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.111466 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:17Z","lastTransitionTime":"2025-12-01T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.171787 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.172011 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:17 crc kubenswrapper[4868]: E1201 17:27:17.172188 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:17 crc kubenswrapper[4868]: E1201 17:27:17.172520 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.215013 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.215360 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.215420 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.215528 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.215597 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:17Z","lastTransitionTime":"2025-12-01T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.319471 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.319556 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.319572 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.319598 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.319614 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:17Z","lastTransitionTime":"2025-12-01T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.423398 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.423452 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.423468 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.423493 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.423511 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:17Z","lastTransitionTime":"2025-12-01T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.527040 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.527130 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.527145 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.527168 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.527183 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:17Z","lastTransitionTime":"2025-12-01T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.631018 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.631073 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.631086 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.631105 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.631117 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:17Z","lastTransitionTime":"2025-12-01T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.734551 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.734659 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.734687 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.734735 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.734767 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:17Z","lastTransitionTime":"2025-12-01T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.837899 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.838028 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.838042 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.838060 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.838079 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:17Z","lastTransitionTime":"2025-12-01T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.940605 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.940648 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.940657 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.940671 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:17 crc kubenswrapper[4868]: I1201 17:27:17.940681 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:17Z","lastTransitionTime":"2025-12-01T17:27:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.044097 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.044159 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.044179 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.044207 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.044226 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:18Z","lastTransitionTime":"2025-12-01T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.147873 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.147990 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.148010 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.148038 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.148061 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:18Z","lastTransitionTime":"2025-12-01T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.171412 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.171527 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:18 crc kubenswrapper[4868]: E1201 17:27:18.171597 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:18 crc kubenswrapper[4868]: E1201 17:27:18.171695 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.229731 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=55.229694951 podStartE2EDuration="55.229694951s" podCreationTimestamp="2025-12-01 17:26:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:18.216777405 +0000 UTC m=+110.587887816" watchObservedRunningTime="2025-12-01 17:27:18.229694951 +0000 UTC m=+110.600805372" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.252916 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.253000 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.253011 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.253029 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.253042 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:18Z","lastTransitionTime":"2025-12-01T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.279193 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=20.279167276 podStartE2EDuration="20.279167276s" podCreationTimestamp="2025-12-01 17:26:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:18.27856041 +0000 UTC m=+110.649670951" watchObservedRunningTime="2025-12-01 17:27:18.279167276 +0000 UTC m=+110.650277687" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.309917 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=91.309896525 podStartE2EDuration="1m31.309896525s" podCreationTimestamp="2025-12-01 17:25:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:18.296771303 +0000 UTC m=+110.667881714" watchObservedRunningTime="2025-12-01 17:27:18.309896525 +0000 UTC m=+110.681006936" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.330149 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c7p72" podStartSLOduration=84.3301147 podStartE2EDuration="1m24.3301147s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:18.321911886 +0000 UTC m=+110.693022297" watchObservedRunningTime="2025-12-01 17:27:18.3301147 +0000 UTC m=+110.701225111" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.355238 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=6.355205222 podStartE2EDuration="6.355205222s" podCreationTimestamp="2025-12-01 17:27:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:18.354705609 +0000 UTC m=+110.725816030" watchObservedRunningTime="2025-12-01 17:27:18.355205222 +0000 UTC m=+110.726315653" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.360516 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.360556 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.360564 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.360580 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.360593 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:18Z","lastTransitionTime":"2025-12-01T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.385790 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-h9tlw" podStartSLOduration=85.385760115 podStartE2EDuration="1m25.385760115s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:18.371625928 +0000 UTC m=+110.742736339" watchObservedRunningTime="2025-12-01 17:27:18.385760115 +0000 UTC m=+110.756870526" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.410762 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podStartSLOduration=85.410734734 podStartE2EDuration="1m25.410734734s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:18.386441393 +0000 UTC m=+110.757551804" watchObservedRunningTime="2025-12-01 17:27:18.410734734 +0000 UTC m=+110.781845145" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.425398 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-64mjk" podStartSLOduration=85.425372524 podStartE2EDuration="1m25.425372524s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:18.424958834 +0000 UTC m=+110.796069245" watchObservedRunningTime="2025-12-01 17:27:18.425372524 +0000 UTC m=+110.796482935" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.448741 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=88.44871293 podStartE2EDuration="1m28.44871293s" podCreationTimestamp="2025-12-01 17:25:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:18.44792573 +0000 UTC m=+110.819036151" watchObservedRunningTime="2025-12-01 17:27:18.44871293 +0000 UTC m=+110.819823351" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.462656 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.462697 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.462708 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.462723 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.462736 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:18Z","lastTransitionTime":"2025-12-01T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.477648 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-c7mwj" podStartSLOduration=85.477627801 podStartE2EDuration="1m25.477627801s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:18.476607945 +0000 UTC m=+110.847718366" watchObservedRunningTime="2025-12-01 17:27:18.477627801 +0000 UTC m=+110.848738212" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.500062 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-hw725" podStartSLOduration=84.500040503 podStartE2EDuration="1m24.500040503s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:18.49873584 +0000 UTC m=+110.869846271" watchObservedRunningTime="2025-12-01 17:27:18.500040503 +0000 UTC m=+110.871150914" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.566031 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.566075 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.566086 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.566104 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.566117 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:18Z","lastTransitionTime":"2025-12-01T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.669167 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.669508 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.669576 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.669697 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.669774 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:18Z","lastTransitionTime":"2025-12-01T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.772193 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.772261 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.772281 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.772310 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.772330 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:18Z","lastTransitionTime":"2025-12-01T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.875333 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.875392 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.875403 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.875427 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.875441 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:18Z","lastTransitionTime":"2025-12-01T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.984546 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.984597 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.984611 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.984637 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:18 crc kubenswrapper[4868]: I1201 17:27:18.984651 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:18Z","lastTransitionTime":"2025-12-01T17:27:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.087666 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.087709 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.087718 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.087734 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.087752 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:19Z","lastTransitionTime":"2025-12-01T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.171851 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.171994 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:19 crc kubenswrapper[4868]: E1201 17:27:19.172489 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:19 crc kubenswrapper[4868]: E1201 17:27:19.172660 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.173185 4868 scope.go:117] "RemoveContainer" containerID="fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c" Dec 01 17:27:19 crc kubenswrapper[4868]: E1201 17:27:19.173339 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-q9z6l_openshift-ovn-kubernetes(5baca705-90cc-4198-b641-214a2b915086)\"" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podUID="5baca705-90cc-4198-b641-214a2b915086" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.189775 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.189814 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.189826 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.189843 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.189854 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:19Z","lastTransitionTime":"2025-12-01T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.292691 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.292742 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.292751 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.292772 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.292787 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:19Z","lastTransitionTime":"2025-12-01T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.401091 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.401202 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.401219 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.401237 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.401249 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:19Z","lastTransitionTime":"2025-12-01T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.504412 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.504705 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.504770 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.504836 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.504911 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:19Z","lastTransitionTime":"2025-12-01T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.608322 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.608362 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.608373 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.608391 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.608402 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:19Z","lastTransitionTime":"2025-12-01T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.712077 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.712156 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.712175 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.712206 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.712228 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:19Z","lastTransitionTime":"2025-12-01T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.816326 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.816401 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.816419 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.816448 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.816468 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:19Z","lastTransitionTime":"2025-12-01T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.919014 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.919312 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.919386 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.919458 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:19 crc kubenswrapper[4868]: I1201 17:27:19.919518 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:19Z","lastTransitionTime":"2025-12-01T17:27:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.022850 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.022879 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.022888 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.022903 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.022914 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:20Z","lastTransitionTime":"2025-12-01T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.126531 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.126574 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.126585 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.126600 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.126610 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:20Z","lastTransitionTime":"2025-12-01T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.172052 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.172145 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:20 crc kubenswrapper[4868]: E1201 17:27:20.172316 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:20 crc kubenswrapper[4868]: E1201 17:27:20.172504 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.229996 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.230045 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.230056 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.230078 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.230090 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:20Z","lastTransitionTime":"2025-12-01T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.333254 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.333302 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.333315 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.333331 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.333343 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:20Z","lastTransitionTime":"2025-12-01T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.437306 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.437370 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.437388 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.437414 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.437431 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:20Z","lastTransitionTime":"2025-12-01T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.541319 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.541379 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.541391 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.541418 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.541432 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:20Z","lastTransitionTime":"2025-12-01T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.644192 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.644279 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.644305 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.644342 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.644368 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:20Z","lastTransitionTime":"2025-12-01T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.748292 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.748350 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.748368 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.748400 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.748420 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:20Z","lastTransitionTime":"2025-12-01T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.851321 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.851405 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.851430 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.851463 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.851489 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:20Z","lastTransitionTime":"2025-12-01T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.955231 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.955289 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.955301 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.955322 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:20 crc kubenswrapper[4868]: I1201 17:27:20.955332 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:20Z","lastTransitionTime":"2025-12-01T17:27:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.059148 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.059202 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.059215 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.059232 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.059245 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:21Z","lastTransitionTime":"2025-12-01T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.163261 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.163363 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.163390 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.163428 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.163452 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:21Z","lastTransitionTime":"2025-12-01T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.171863 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:21 crc kubenswrapper[4868]: E1201 17:27:21.172096 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.172367 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:21 crc kubenswrapper[4868]: E1201 17:27:21.172635 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.266527 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.266583 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.266600 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.266619 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.266632 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:21Z","lastTransitionTime":"2025-12-01T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.369178 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.369277 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.369301 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.369336 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.369359 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:21Z","lastTransitionTime":"2025-12-01T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.472137 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.472186 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.472199 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.472225 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.472238 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:21Z","lastTransitionTime":"2025-12-01T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.575555 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.576127 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.576149 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.576180 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.576201 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:21Z","lastTransitionTime":"2025-12-01T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.679158 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.679215 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.679228 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.679247 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.679261 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:21Z","lastTransitionTime":"2025-12-01T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.782578 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.782626 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.782639 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.782659 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.782670 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:21Z","lastTransitionTime":"2025-12-01T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.886807 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.886895 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.886924 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.887002 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.887026 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:21Z","lastTransitionTime":"2025-12-01T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.989562 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.990026 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.990135 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.990210 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:21 crc kubenswrapper[4868]: I1201 17:27:21.990273 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:21Z","lastTransitionTime":"2025-12-01T17:27:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.093912 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.093994 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.094006 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.094023 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.094035 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:22Z","lastTransitionTime":"2025-12-01T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.171513 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:22 crc kubenswrapper[4868]: E1201 17:27:22.171742 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.171955 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:22 crc kubenswrapper[4868]: E1201 17:27:22.172151 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.197524 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.197595 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.197607 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.197629 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.197644 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:22Z","lastTransitionTime":"2025-12-01T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.305882 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.306853 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.306873 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.306971 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.307005 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:22Z","lastTransitionTime":"2025-12-01T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.410891 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.410966 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.410978 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.411002 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.411017 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:22Z","lastTransitionTime":"2025-12-01T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.514608 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.514679 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.514693 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.514712 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.514721 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:22Z","lastTransitionTime":"2025-12-01T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.618610 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.618668 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.618685 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.618707 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.618722 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:22Z","lastTransitionTime":"2025-12-01T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.721337 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.721378 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.721393 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.721440 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.721456 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:22Z","lastTransitionTime":"2025-12-01T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.825064 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.825120 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.825132 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.825152 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.825166 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:22Z","lastTransitionTime":"2025-12-01T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.928129 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.928183 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.928193 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.928210 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:22 crc kubenswrapper[4868]: I1201 17:27:22.928223 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:22Z","lastTransitionTime":"2025-12-01T17:27:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.030957 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.031005 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.031017 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.031035 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.031045 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:23Z","lastTransitionTime":"2025-12-01T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.135242 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.135318 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.135340 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.135370 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.135391 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:23Z","lastTransitionTime":"2025-12-01T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.171268 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:23 crc kubenswrapper[4868]: E1201 17:27:23.171414 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.171279 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:23 crc kubenswrapper[4868]: E1201 17:27:23.171494 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.238418 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.238542 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.238577 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.238618 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.238651 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:23Z","lastTransitionTime":"2025-12-01T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.341109 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.341189 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.341211 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.341241 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.341260 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:23Z","lastTransitionTime":"2025-12-01T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.444324 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.444402 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.444420 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.444448 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.444469 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:23Z","lastTransitionTime":"2025-12-01T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.547466 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.547553 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.547573 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.547605 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.547625 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:23Z","lastTransitionTime":"2025-12-01T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.650617 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.650669 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.650686 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.650709 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.650721 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:23Z","lastTransitionTime":"2025-12-01T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.753012 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.753118 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.753135 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.753164 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.753182 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:23Z","lastTransitionTime":"2025-12-01T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.856109 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.856143 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.856153 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.856168 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.856179 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:23Z","lastTransitionTime":"2025-12-01T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.959057 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.959102 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.959115 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.959133 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:23 crc kubenswrapper[4868]: I1201 17:27:23.959143 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:23Z","lastTransitionTime":"2025-12-01T17:27:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.061112 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.061157 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.061165 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.061185 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.061197 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:24Z","lastTransitionTime":"2025-12-01T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.163548 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.163596 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.163610 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.163630 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.163647 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:24Z","lastTransitionTime":"2025-12-01T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.171931 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:24 crc kubenswrapper[4868]: E1201 17:27:24.172104 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.172157 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:24 crc kubenswrapper[4868]: E1201 17:27:24.172367 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.267601 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.267658 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.267670 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.267689 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.267701 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:24Z","lastTransitionTime":"2025-12-01T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.365151 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.365196 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.365214 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.365234 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.365246 4868 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T17:27:24Z","lastTransitionTime":"2025-12-01T17:27:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.411130 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx"] Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.411619 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.415363 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.415417 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.415701 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.415763 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.477235 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/39979b11-fa94-4d5a-a63c-ac986aee5468-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5r6gx\" (UID: \"39979b11-fa94-4d5a-a63c-ac986aee5468\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.477328 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/39979b11-fa94-4d5a-a63c-ac986aee5468-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5r6gx\" (UID: \"39979b11-fa94-4d5a-a63c-ac986aee5468\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.477351 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/39979b11-fa94-4d5a-a63c-ac986aee5468-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5r6gx\" (UID: \"39979b11-fa94-4d5a-a63c-ac986aee5468\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.477381 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/39979b11-fa94-4d5a-a63c-ac986aee5468-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5r6gx\" (UID: \"39979b11-fa94-4d5a-a63c-ac986aee5468\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.477408 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39979b11-fa94-4d5a-a63c-ac986aee5468-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5r6gx\" (UID: \"39979b11-fa94-4d5a-a63c-ac986aee5468\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.578403 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/39979b11-fa94-4d5a-a63c-ac986aee5468-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5r6gx\" (UID: \"39979b11-fa94-4d5a-a63c-ac986aee5468\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.578487 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/39979b11-fa94-4d5a-a63c-ac986aee5468-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5r6gx\" (UID: \"39979b11-fa94-4d5a-a63c-ac986aee5468\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.578519 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/39979b11-fa94-4d5a-a63c-ac986aee5468-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5r6gx\" (UID: \"39979b11-fa94-4d5a-a63c-ac986aee5468\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.578544 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/39979b11-fa94-4d5a-a63c-ac986aee5468-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5r6gx\" (UID: \"39979b11-fa94-4d5a-a63c-ac986aee5468\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.578571 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39979b11-fa94-4d5a-a63c-ac986aee5468-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5r6gx\" (UID: \"39979b11-fa94-4d5a-a63c-ac986aee5468\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.578758 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/39979b11-fa94-4d5a-a63c-ac986aee5468-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5r6gx\" (UID: \"39979b11-fa94-4d5a-a63c-ac986aee5468\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.578912 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/39979b11-fa94-4d5a-a63c-ac986aee5468-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5r6gx\" (UID: \"39979b11-fa94-4d5a-a63c-ac986aee5468\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.579742 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/39979b11-fa94-4d5a-a63c-ac986aee5468-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5r6gx\" (UID: \"39979b11-fa94-4d5a-a63c-ac986aee5468\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.584614 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39979b11-fa94-4d5a-a63c-ac986aee5468-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5r6gx\" (UID: \"39979b11-fa94-4d5a-a63c-ac986aee5468\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.597593 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/39979b11-fa94-4d5a-a63c-ac986aee5468-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5r6gx\" (UID: \"39979b11-fa94-4d5a-a63c-ac986aee5468\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.727054 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" Dec 01 17:27:24 crc kubenswrapper[4868]: I1201 17:27:24.856066 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" event={"ID":"39979b11-fa94-4d5a-a63c-ac986aee5468","Type":"ContainerStarted","Data":"f3226ddb726a5579e6a52c81ae1d3a997eb358557d562102bf89387261abc99c"} Dec 01 17:27:25 crc kubenswrapper[4868]: I1201 17:27:25.171418 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:25 crc kubenswrapper[4868]: I1201 17:27:25.171468 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:25 crc kubenswrapper[4868]: E1201 17:27:25.171595 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:25 crc kubenswrapper[4868]: E1201 17:27:25.171782 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:25 crc kubenswrapper[4868]: I1201 17:27:25.865592 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" event={"ID":"39979b11-fa94-4d5a-a63c-ac986aee5468","Type":"ContainerStarted","Data":"40343b21875ffa426d5ddd80a17b0a012aecc5e970f613ee05b5af69325054f0"} Dec 01 17:27:25 crc kubenswrapper[4868]: I1201 17:27:25.882452 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5r6gx" podStartSLOduration=92.882435438 podStartE2EDuration="1m32.882435438s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:25.881973396 +0000 UTC m=+118.253083807" watchObservedRunningTime="2025-12-01 17:27:25.882435438 +0000 UTC m=+118.253545849" Dec 01 17:27:26 crc kubenswrapper[4868]: I1201 17:27:26.170905 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:26 crc kubenswrapper[4868]: I1201 17:27:26.171009 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:26 crc kubenswrapper[4868]: E1201 17:27:26.171096 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:26 crc kubenswrapper[4868]: E1201 17:27:26.171153 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:27 crc kubenswrapper[4868]: I1201 17:27:27.171245 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:27 crc kubenswrapper[4868]: I1201 17:27:27.171271 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:27 crc kubenswrapper[4868]: E1201 17:27:27.171771 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:27 crc kubenswrapper[4868]: E1201 17:27:27.171897 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:28 crc kubenswrapper[4868]: E1201 17:27:28.099368 4868 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 01 17:27:28 crc kubenswrapper[4868]: I1201 17:27:28.171238 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:28 crc kubenswrapper[4868]: I1201 17:27:28.171304 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:28 crc kubenswrapper[4868]: E1201 17:27:28.173700 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:28 crc kubenswrapper[4868]: E1201 17:27:28.174004 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:28 crc kubenswrapper[4868]: E1201 17:27:28.294096 4868 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 17:27:29 crc kubenswrapper[4868]: I1201 17:27:29.182263 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:29 crc kubenswrapper[4868]: I1201 17:27:29.182338 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:29 crc kubenswrapper[4868]: E1201 17:27:29.182448 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:29 crc kubenswrapper[4868]: E1201 17:27:29.182716 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:30 crc kubenswrapper[4868]: I1201 17:27:30.171886 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:30 crc kubenswrapper[4868]: I1201 17:27:30.172169 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:30 crc kubenswrapper[4868]: E1201 17:27:30.172367 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:30 crc kubenswrapper[4868]: E1201 17:27:30.172877 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:30 crc kubenswrapper[4868]: I1201 17:27:30.173249 4868 scope.go:117] "RemoveContainer" containerID="fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c" Dec 01 17:27:30 crc kubenswrapper[4868]: E1201 17:27:30.173437 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-q9z6l_openshift-ovn-kubernetes(5baca705-90cc-4198-b641-214a2b915086)\"" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podUID="5baca705-90cc-4198-b641-214a2b915086" Dec 01 17:27:31 crc kubenswrapper[4868]: I1201 17:27:31.171373 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:31 crc kubenswrapper[4868]: I1201 17:27:31.171436 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:31 crc kubenswrapper[4868]: E1201 17:27:31.171579 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:31 crc kubenswrapper[4868]: E1201 17:27:31.171777 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:31 crc kubenswrapper[4868]: I1201 17:27:31.891910 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9tlw_8ce688b4-36b1-48d4-a6d1-8aec723125c4/kube-multus/1.log" Dec 01 17:27:31 crc kubenswrapper[4868]: I1201 17:27:31.892916 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9tlw_8ce688b4-36b1-48d4-a6d1-8aec723125c4/kube-multus/0.log" Dec 01 17:27:31 crc kubenswrapper[4868]: I1201 17:27:31.893014 4868 generic.go:334] "Generic (PLEG): container finished" podID="8ce688b4-36b1-48d4-a6d1-8aec723125c4" containerID="3d8b1e3f6f1f6f8ed2844b1e533905130d255430f7031699ce02dfa6a3ac16a5" exitCode=1 Dec 01 17:27:31 crc kubenswrapper[4868]: I1201 17:27:31.893059 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9tlw" event={"ID":"8ce688b4-36b1-48d4-a6d1-8aec723125c4","Type":"ContainerDied","Data":"3d8b1e3f6f1f6f8ed2844b1e533905130d255430f7031699ce02dfa6a3ac16a5"} Dec 01 17:27:31 crc kubenswrapper[4868]: I1201 17:27:31.893143 4868 scope.go:117] "RemoveContainer" containerID="b56f0ae0cfb81948e738b4aa115385719a51afbec453c6338f8aabe6c81b0c58" Dec 01 17:27:31 crc kubenswrapper[4868]: I1201 17:27:31.893777 4868 scope.go:117] "RemoveContainer" containerID="3d8b1e3f6f1f6f8ed2844b1e533905130d255430f7031699ce02dfa6a3ac16a5" Dec 01 17:27:31 crc kubenswrapper[4868]: E1201 17:27:31.894218 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-h9tlw_openshift-multus(8ce688b4-36b1-48d4-a6d1-8aec723125c4)\"" pod="openshift-multus/multus-h9tlw" podUID="8ce688b4-36b1-48d4-a6d1-8aec723125c4" Dec 01 17:27:32 crc kubenswrapper[4868]: I1201 17:27:32.171761 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:32 crc kubenswrapper[4868]: I1201 17:27:32.171762 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:32 crc kubenswrapper[4868]: E1201 17:27:32.172003 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:32 crc kubenswrapper[4868]: E1201 17:27:32.172126 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:32 crc kubenswrapper[4868]: I1201 17:27:32.898796 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9tlw_8ce688b4-36b1-48d4-a6d1-8aec723125c4/kube-multus/1.log" Dec 01 17:27:33 crc kubenswrapper[4868]: I1201 17:27:33.171402 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:33 crc kubenswrapper[4868]: I1201 17:27:33.171548 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:33 crc kubenswrapper[4868]: E1201 17:27:33.171693 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:33 crc kubenswrapper[4868]: E1201 17:27:33.171935 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:33 crc kubenswrapper[4868]: E1201 17:27:33.296200 4868 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 17:27:34 crc kubenswrapper[4868]: I1201 17:27:34.171915 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:34 crc kubenswrapper[4868]: E1201 17:27:34.172094 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:34 crc kubenswrapper[4868]: I1201 17:27:34.172588 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:34 crc kubenswrapper[4868]: E1201 17:27:34.172804 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:35 crc kubenswrapper[4868]: I1201 17:27:35.171891 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:35 crc kubenswrapper[4868]: E1201 17:27:35.172150 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:35 crc kubenswrapper[4868]: I1201 17:27:35.173010 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:35 crc kubenswrapper[4868]: E1201 17:27:35.173309 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:36 crc kubenswrapper[4868]: I1201 17:27:36.171371 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:36 crc kubenswrapper[4868]: I1201 17:27:36.171510 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:36 crc kubenswrapper[4868]: E1201 17:27:36.171911 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:36 crc kubenswrapper[4868]: E1201 17:27:36.172243 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:37 crc kubenswrapper[4868]: I1201 17:27:37.171375 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:37 crc kubenswrapper[4868]: I1201 17:27:37.171442 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:37 crc kubenswrapper[4868]: E1201 17:27:37.171534 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:37 crc kubenswrapper[4868]: E1201 17:27:37.171759 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:38 crc kubenswrapper[4868]: I1201 17:27:38.175050 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:38 crc kubenswrapper[4868]: I1201 17:27:38.175630 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:38 crc kubenswrapper[4868]: E1201 17:27:38.175957 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:38 crc kubenswrapper[4868]: E1201 17:27:38.175967 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:38 crc kubenswrapper[4868]: E1201 17:27:38.298072 4868 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 17:27:39 crc kubenswrapper[4868]: I1201 17:27:39.171663 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:39 crc kubenswrapper[4868]: I1201 17:27:39.171736 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:39 crc kubenswrapper[4868]: E1201 17:27:39.171811 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:39 crc kubenswrapper[4868]: E1201 17:27:39.171853 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:40 crc kubenswrapper[4868]: I1201 17:27:40.171825 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:40 crc kubenswrapper[4868]: I1201 17:27:40.171979 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:40 crc kubenswrapper[4868]: E1201 17:27:40.172128 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:40 crc kubenswrapper[4868]: E1201 17:27:40.172376 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:41 crc kubenswrapper[4868]: I1201 17:27:41.171068 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:41 crc kubenswrapper[4868]: I1201 17:27:41.171474 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:41 crc kubenswrapper[4868]: E1201 17:27:41.171540 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:41 crc kubenswrapper[4868]: E1201 17:27:41.171786 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:41 crc kubenswrapper[4868]: I1201 17:27:41.171812 4868 scope.go:117] "RemoveContainer" containerID="fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c" Dec 01 17:27:41 crc kubenswrapper[4868]: I1201 17:27:41.934608 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovnkube-controller/3.log" Dec 01 17:27:41 crc kubenswrapper[4868]: I1201 17:27:41.937470 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerStarted","Data":"ede2d062bb6d3ec0c91e1b108217d9466e8941a233d7e8f2e8c18f033490fb99"} Dec 01 17:27:41 crc kubenswrapper[4868]: I1201 17:27:41.938608 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:27:42 crc kubenswrapper[4868]: I1201 17:27:42.172024 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:42 crc kubenswrapper[4868]: I1201 17:27:42.172109 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:42 crc kubenswrapper[4868]: E1201 17:27:42.172290 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:42 crc kubenswrapper[4868]: E1201 17:27:42.172401 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:42 crc kubenswrapper[4868]: I1201 17:27:42.584672 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podStartSLOduration=108.58464075 podStartE2EDuration="1m48.58464075s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:41.996552883 +0000 UTC m=+134.367663294" watchObservedRunningTime="2025-12-01 17:27:42.58464075 +0000 UTC m=+134.955751161" Dec 01 17:27:42 crc kubenswrapper[4868]: I1201 17:27:42.586442 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-nxfkx"] Dec 01 17:27:42 crc kubenswrapper[4868]: I1201 17:27:42.947525 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:42 crc kubenswrapper[4868]: E1201 17:27:42.947737 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:43 crc kubenswrapper[4868]: I1201 17:27:43.172417 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:43 crc kubenswrapper[4868]: E1201 17:27:43.172531 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:43 crc kubenswrapper[4868]: I1201 17:27:43.172767 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:43 crc kubenswrapper[4868]: E1201 17:27:43.172829 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:43 crc kubenswrapper[4868]: I1201 17:27:43.173174 4868 scope.go:117] "RemoveContainer" containerID="3d8b1e3f6f1f6f8ed2844b1e533905130d255430f7031699ce02dfa6a3ac16a5" Dec 01 17:27:43 crc kubenswrapper[4868]: E1201 17:27:43.299830 4868 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 17:27:43 crc kubenswrapper[4868]: I1201 17:27:43.953394 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9tlw_8ce688b4-36b1-48d4-a6d1-8aec723125c4/kube-multus/1.log" Dec 01 17:27:43 crc kubenswrapper[4868]: I1201 17:27:43.953461 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9tlw" event={"ID":"8ce688b4-36b1-48d4-a6d1-8aec723125c4","Type":"ContainerStarted","Data":"54e1c435963c9a72f109a2d0e042ba4b27e7e7544588ebfff04e4c7042eb5335"} Dec 01 17:27:44 crc kubenswrapper[4868]: I1201 17:27:44.171081 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:44 crc kubenswrapper[4868]: E1201 17:27:44.171352 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:44 crc kubenswrapper[4868]: I1201 17:27:44.171481 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:44 crc kubenswrapper[4868]: E1201 17:27:44.171736 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:45 crc kubenswrapper[4868]: I1201 17:27:45.171816 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:45 crc kubenswrapper[4868]: I1201 17:27:45.171856 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:45 crc kubenswrapper[4868]: E1201 17:27:45.172110 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:45 crc kubenswrapper[4868]: E1201 17:27:45.172273 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:46 crc kubenswrapper[4868]: I1201 17:27:46.170843 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:46 crc kubenswrapper[4868]: I1201 17:27:46.170896 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:46 crc kubenswrapper[4868]: E1201 17:27:46.171080 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:46 crc kubenswrapper[4868]: E1201 17:27:46.171211 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:47 crc kubenswrapper[4868]: I1201 17:27:47.171028 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:47 crc kubenswrapper[4868]: I1201 17:27:47.171054 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:47 crc kubenswrapper[4868]: E1201 17:27:47.171224 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 17:27:47 crc kubenswrapper[4868]: E1201 17:27:47.171328 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 17:27:48 crc kubenswrapper[4868]: I1201 17:27:48.171870 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:48 crc kubenswrapper[4868]: I1201 17:27:48.172036 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:48 crc kubenswrapper[4868]: E1201 17:27:48.172829 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nxfkx" podUID="4afd9a0c-e8c7-42ec-9771-036c3441cbce" Dec 01 17:27:48 crc kubenswrapper[4868]: E1201 17:27:48.173020 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 17:27:49 crc kubenswrapper[4868]: I1201 17:27:49.171762 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:49 crc kubenswrapper[4868]: I1201 17:27:49.171816 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:49 crc kubenswrapper[4868]: I1201 17:27:49.174111 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 17:27:49 crc kubenswrapper[4868]: I1201 17:27:49.174158 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 17:27:49 crc kubenswrapper[4868]: I1201 17:27:49.174191 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 17:27:49 crc kubenswrapper[4868]: I1201 17:27:49.174350 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 17:27:50 crc kubenswrapper[4868]: I1201 17:27:50.171315 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:27:50 crc kubenswrapper[4868]: I1201 17:27:50.171333 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:50 crc kubenswrapper[4868]: I1201 17:27:50.175020 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 17:27:50 crc kubenswrapper[4868]: I1201 17:27:50.175306 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 17:27:52 crc kubenswrapper[4868]: I1201 17:27:52.408180 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.029578 4868 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.060809 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qw86w"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.061651 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.061823 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hwgjb"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.062623 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.062735 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pvqvn"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.063525 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.065820 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.066907 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.067267 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.067342 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.067665 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.067808 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.067920 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.068040 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.068266 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.069203 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.069313 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.069404 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.069680 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.069829 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.070202 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.070263 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.070813 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.070893 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.071414 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.072227 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.072859 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.073755 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.073991 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.074558 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.074802 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.075272 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.075448 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vr48t"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.076145 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.076209 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-grsv4"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.076839 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.076964 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.077009 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.077190 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.077543 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.078117 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.078426 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.079077 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.079139 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.079348 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.079807 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-s4hdj"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.080118 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.080213 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.080280 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.080384 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.080842 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.081215 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.082859 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.083034 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.083402 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.083968 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.084229 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.094083 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.096005 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.102281 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.103197 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.103366 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vr48t"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.103644 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.106616 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hwgjb"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.106934 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.108768 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.109053 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.109210 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.109308 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.109479 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.110033 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.110245 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.113196 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.113327 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.113383 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.113546 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.113652 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.113731 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.114317 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.114431 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.114457 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.114538 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.114703 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.113552 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.115032 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.115134 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.115183 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.115238 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.115272 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.115335 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.115342 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.115464 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.115536 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qw86w"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.116471 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.116824 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.116880 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.119309 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.121899 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-grsv4"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.124627 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129451 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129598 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/be098182-dd51-4a19-920a-a797a175124a-images\") pod \"machine-api-operator-5694c8668f-qw86w\" (UID: \"be098182-dd51-4a19-920a-a797a175124a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129619 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-config\") pod \"controller-manager-879f6c89f-s4hdj\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129637 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129659 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/221dbb7a-96df-41b4-9c28-240d63013e93-node-pullsecrets\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129677 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129696 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hftzc\" (UniqueName: \"kubernetes.io/projected/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-kube-api-access-hftzc\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129713 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abf290ff-6ae2-4354-a3df-7103c94e481e-config\") pod \"machine-approver-56656f9798-qcjvq\" (UID: \"abf290ff-6ae2-4354-a3df-7103c94e481e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129727 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-audit-dir\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129744 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/221dbb7a-96df-41b4-9c28-240d63013e93-audit-dir\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129772 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-audit-policies\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129790 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129815 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/221dbb7a-96df-41b4-9c28-240d63013e93-etcd-serving-ca\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129833 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/221dbb7a-96df-41b4-9c28-240d63013e93-encryption-config\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129850 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe918f77-d000-44a8-9b16-c106f62cdede-config\") pod \"authentication-operator-69f744f599-pvqvn\" (UID: \"fe918f77-d000-44a8-9b16-c106f62cdede\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129864 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129880 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2de90211-84d3-45fc-87eb-efb7d7f16a33-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vr48t\" (UID: \"2de90211-84d3-45fc-87eb-efb7d7f16a33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129897 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/221dbb7a-96df-41b4-9c28-240d63013e93-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129914 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjhd6\" (UniqueName: \"kubernetes.io/projected/be098182-dd51-4a19-920a-a797a175124a-kube-api-access-hjhd6\") pod \"machine-api-operator-5694c8668f-qw86w\" (UID: \"be098182-dd51-4a19-920a-a797a175124a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.129931 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe918f77-d000-44a8-9b16-c106f62cdede-serving-cert\") pod \"authentication-operator-69f744f599-pvqvn\" (UID: \"fe918f77-d000-44a8-9b16-c106f62cdede\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130010 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhcwr\" (UniqueName: \"kubernetes.io/projected/abf290ff-6ae2-4354-a3df-7103c94e481e-kube-api-access-xhcwr\") pod \"machine-approver-56656f9798-qcjvq\" (UID: \"abf290ff-6ae2-4354-a3df-7103c94e481e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130028 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-audit-dir\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130045 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130063 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130085 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130149 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9prz\" (UniqueName: \"kubernetes.io/projected/fe918f77-d000-44a8-9b16-c106f62cdede-kube-api-access-w9prz\") pod \"authentication-operator-69f744f599-pvqvn\" (UID: \"fe918f77-d000-44a8-9b16-c106f62cdede\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130166 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130181 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/221dbb7a-96df-41b4-9c28-240d63013e93-etcd-client\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130199 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be098182-dd51-4a19-920a-a797a175124a-config\") pod \"machine-api-operator-5694c8668f-qw86w\" (UID: \"be098182-dd51-4a19-920a-a797a175124a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130224 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/be098182-dd51-4a19-920a-a797a175124a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qw86w\" (UID: \"be098182-dd51-4a19-920a-a797a175124a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130240 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tjdp\" (UniqueName: \"kubernetes.io/projected/b124c538-2e03-45ea-9666-b52b1fc90bdb-kube-api-access-2tjdp\") pod \"controller-manager-879f6c89f-s4hdj\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130263 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130302 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-client-ca\") pod \"controller-manager-879f6c89f-s4hdj\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130326 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130343 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7kxt\" (UniqueName: \"kubernetes.io/projected/baaddd6b-cbef-451e-9315-2e64908dfaa2-kube-api-access-d7kxt\") pod \"openshift-apiserver-operator-796bbdcf4f-48kjd\" (UID: \"baaddd6b-cbef-451e-9315-2e64908dfaa2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130358 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe918f77-d000-44a8-9b16-c106f62cdede-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pvqvn\" (UID: \"fe918f77-d000-44a8-9b16-c106f62cdede\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130378 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130392 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-serving-cert\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130429 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/abf290ff-6ae2-4354-a3df-7103c94e481e-auth-proxy-config\") pod \"machine-approver-56656f9798-qcjvq\" (UID: \"abf290ff-6ae2-4354-a3df-7103c94e481e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130468 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-etcd-client\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130484 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzd2r\" (UniqueName: \"kubernetes.io/projected/2de90211-84d3-45fc-87eb-efb7d7f16a33-kube-api-access-vzd2r\") pod \"openshift-config-operator-7777fb866f-vr48t\" (UID: \"2de90211-84d3-45fc-87eb-efb7d7f16a33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130499 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/221dbb7a-96df-41b4-9c28-240d63013e93-image-import-ca\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130515 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b124c538-2e03-45ea-9666-b52b1fc90bdb-serving-cert\") pod \"controller-manager-879f6c89f-s4hdj\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130528 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/baaddd6b-cbef-451e-9315-2e64908dfaa2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-48kjd\" (UID: \"baaddd6b-cbef-451e-9315-2e64908dfaa2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130543 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe918f77-d000-44a8-9b16-c106f62cdede-service-ca-bundle\") pod \"authentication-operator-69f744f599-pvqvn\" (UID: \"fe918f77-d000-44a8-9b16-c106f62cdede\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130559 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcv9g\" (UniqueName: \"kubernetes.io/projected/a743581d-083b-4400-a1f3-a73f1cfece31-kube-api-access-jcv9g\") pod \"route-controller-manager-6576b87f9c-qc4x6\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130582 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/baaddd6b-cbef-451e-9315-2e64908dfaa2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-48kjd\" (UID: \"baaddd6b-cbef-451e-9315-2e64908dfaa2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130597 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-encryption-config\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130637 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/221dbb7a-96df-41b4-9c28-240d63013e93-config\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130653 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/221dbb7a-96df-41b4-9c28-240d63013e93-audit\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130666 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/221dbb7a-96df-41b4-9c28-240d63013e93-serving-cert\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130686 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130705 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-audit-policies\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130721 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppv8z\" (UniqueName: \"kubernetes.io/projected/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-kube-api-access-ppv8z\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130737 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130752 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/abf290ff-6ae2-4354-a3df-7103c94e481e-machine-approver-tls\") pod \"machine-approver-56656f9798-qcjvq\" (UID: \"abf290ff-6ae2-4354-a3df-7103c94e481e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130797 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a743581d-083b-4400-a1f3-a73f1cfece31-client-ca\") pod \"route-controller-manager-6576b87f9c-qc4x6\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130815 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130831 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130847 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a743581d-083b-4400-a1f3-a73f1cfece31-config\") pod \"route-controller-manager-6576b87f9c-qc4x6\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130862 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-789lm\" (UniqueName: \"kubernetes.io/projected/221dbb7a-96df-41b4-9c28-240d63013e93-kube-api-access-789lm\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130877 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a743581d-083b-4400-a1f3-a73f1cfece31-serving-cert\") pod \"route-controller-manager-6576b87f9c-qc4x6\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130892 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-s4hdj\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.130907 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2de90211-84d3-45fc-87eb-efb7d7f16a33-serving-cert\") pod \"openshift-config-operator-7777fb866f-vr48t\" (UID: \"2de90211-84d3-45fc-87eb-efb7d7f16a33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" Dec 01 17:27:55 crc kubenswrapper[4868]: E1201 17:27:55.131033 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:29:57.131013749 +0000 UTC m=+269.502124160 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.133520 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.134713 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.135114 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.136332 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.231718 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a743581d-083b-4400-a1f3-a73f1cfece31-config\") pod \"route-controller-manager-6576b87f9c-qc4x6\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.231759 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-789lm\" (UniqueName: \"kubernetes.io/projected/221dbb7a-96df-41b4-9c28-240d63013e93-kube-api-access-789lm\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.231800 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-s4hdj\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.231821 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2de90211-84d3-45fc-87eb-efb7d7f16a33-serving-cert\") pod \"openshift-config-operator-7777fb866f-vr48t\" (UID: \"2de90211-84d3-45fc-87eb-efb7d7f16a33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.231839 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a743581d-083b-4400-a1f3-a73f1cfece31-serving-cert\") pod \"route-controller-manager-6576b87f9c-qc4x6\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.231865 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/be098182-dd51-4a19-920a-a797a175124a-images\") pod \"machine-api-operator-5694c8668f-qw86w\" (UID: \"be098182-dd51-4a19-920a-a797a175124a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.231882 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-config\") pod \"controller-manager-879f6c89f-s4hdj\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.231898 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.231914 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/221dbb7a-96df-41b4-9c28-240d63013e93-node-pullsecrets\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232038 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hftzc\" (UniqueName: \"kubernetes.io/projected/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-kube-api-access-hftzc\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232062 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232083 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abf290ff-6ae2-4354-a3df-7103c94e481e-config\") pod \"machine-approver-56656f9798-qcjvq\" (UID: \"abf290ff-6ae2-4354-a3df-7103c94e481e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232101 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-audit-dir\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232117 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/221dbb7a-96df-41b4-9c28-240d63013e93-audit-dir\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232159 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-audit-policies\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232175 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232192 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/221dbb7a-96df-41b4-9c28-240d63013e93-etcd-serving-ca\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232210 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/221dbb7a-96df-41b4-9c28-240d63013e93-encryption-config\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232229 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe918f77-d000-44a8-9b16-c106f62cdede-config\") pod \"authentication-operator-69f744f599-pvqvn\" (UID: \"fe918f77-d000-44a8-9b16-c106f62cdede\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232246 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232262 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2de90211-84d3-45fc-87eb-efb7d7f16a33-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vr48t\" (UID: \"2de90211-84d3-45fc-87eb-efb7d7f16a33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232277 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/221dbb7a-96df-41b4-9c28-240d63013e93-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232320 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjhd6\" (UniqueName: \"kubernetes.io/projected/be098182-dd51-4a19-920a-a797a175124a-kube-api-access-hjhd6\") pod \"machine-api-operator-5694c8668f-qw86w\" (UID: \"be098182-dd51-4a19-920a-a797a175124a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232338 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe918f77-d000-44a8-9b16-c106f62cdede-serving-cert\") pod \"authentication-operator-69f744f599-pvqvn\" (UID: \"fe918f77-d000-44a8-9b16-c106f62cdede\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232357 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhcwr\" (UniqueName: \"kubernetes.io/projected/abf290ff-6ae2-4354-a3df-7103c94e481e-kube-api-access-xhcwr\") pod \"machine-approver-56656f9798-qcjvq\" (UID: \"abf290ff-6ae2-4354-a3df-7103c94e481e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232375 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-audit-dir\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232393 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232412 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232432 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232474 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/221dbb7a-96df-41b4-9c28-240d63013e93-etcd-client\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232500 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9prz\" (UniqueName: \"kubernetes.io/projected/fe918f77-d000-44a8-9b16-c106f62cdede-kube-api-access-w9prz\") pod \"authentication-operator-69f744f599-pvqvn\" (UID: \"fe918f77-d000-44a8-9b16-c106f62cdede\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232516 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be098182-dd51-4a19-920a-a797a175124a-config\") pod \"machine-api-operator-5694c8668f-qw86w\" (UID: \"be098182-dd51-4a19-920a-a797a175124a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232542 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/be098182-dd51-4a19-920a-a797a175124a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qw86w\" (UID: \"be098182-dd51-4a19-920a-a797a175124a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232560 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tjdp\" (UniqueName: \"kubernetes.io/projected/b124c538-2e03-45ea-9666-b52b1fc90bdb-kube-api-access-2tjdp\") pod \"controller-manager-879f6c89f-s4hdj\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232578 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232596 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-client-ca\") pod \"controller-manager-879f6c89f-s4hdj\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232636 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232665 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7kxt\" (UniqueName: \"kubernetes.io/projected/baaddd6b-cbef-451e-9315-2e64908dfaa2-kube-api-access-d7kxt\") pod \"openshift-apiserver-operator-796bbdcf4f-48kjd\" (UID: \"baaddd6b-cbef-451e-9315-2e64908dfaa2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232681 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe918f77-d000-44a8-9b16-c106f62cdede-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pvqvn\" (UID: \"fe918f77-d000-44a8-9b16-c106f62cdede\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232697 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232714 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-serving-cert\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232729 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/abf290ff-6ae2-4354-a3df-7103c94e481e-auth-proxy-config\") pod \"machine-approver-56656f9798-qcjvq\" (UID: \"abf290ff-6ae2-4354-a3df-7103c94e481e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232743 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-etcd-client\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232759 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzd2r\" (UniqueName: \"kubernetes.io/projected/2de90211-84d3-45fc-87eb-efb7d7f16a33-kube-api-access-vzd2r\") pod \"openshift-config-operator-7777fb866f-vr48t\" (UID: \"2de90211-84d3-45fc-87eb-efb7d7f16a33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232798 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/221dbb7a-96df-41b4-9c28-240d63013e93-image-import-ca\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232814 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/baaddd6b-cbef-451e-9315-2e64908dfaa2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-48kjd\" (UID: \"baaddd6b-cbef-451e-9315-2e64908dfaa2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232832 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b124c538-2e03-45ea-9666-b52b1fc90bdb-serving-cert\") pod \"controller-manager-879f6c89f-s4hdj\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232852 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe918f77-d000-44a8-9b16-c106f62cdede-service-ca-bundle\") pod \"authentication-operator-69f744f599-pvqvn\" (UID: \"fe918f77-d000-44a8-9b16-c106f62cdede\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232873 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcv9g\" (UniqueName: \"kubernetes.io/projected/a743581d-083b-4400-a1f3-a73f1cfece31-kube-api-access-jcv9g\") pod \"route-controller-manager-6576b87f9c-qc4x6\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232889 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/baaddd6b-cbef-451e-9315-2e64908dfaa2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-48kjd\" (UID: \"baaddd6b-cbef-451e-9315-2e64908dfaa2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232903 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-encryption-config\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232919 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/221dbb7a-96df-41b4-9c28-240d63013e93-config\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232981 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.232996 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/221dbb7a-96df-41b4-9c28-240d63013e93-audit\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.233013 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/221dbb7a-96df-41b4-9c28-240d63013e93-serving-cert\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.233030 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.233047 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-audit-policies\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.233064 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppv8z\" (UniqueName: \"kubernetes.io/projected/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-kube-api-access-ppv8z\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.233106 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.233126 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/abf290ff-6ae2-4354-a3df-7103c94e481e-machine-approver-tls\") pod \"machine-approver-56656f9798-qcjvq\" (UID: \"abf290ff-6ae2-4354-a3df-7103c94e481e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.233142 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a743581d-083b-4400-a1f3-a73f1cfece31-client-ca\") pod \"route-controller-manager-6576b87f9c-qc4x6\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.233158 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.233179 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.233308 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/be098182-dd51-4a19-920a-a797a175124a-images\") pod \"machine-api-operator-5694c8668f-qw86w\" (UID: \"be098182-dd51-4a19-920a-a797a175124a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.233554 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a743581d-083b-4400-a1f3-a73f1cfece31-config\") pod \"route-controller-manager-6576b87f9c-qc4x6\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.233637 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-config\") pod \"controller-manager-879f6c89f-s4hdj\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.233701 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-s4hdj\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.233721 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-audit-dir\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.234047 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.234212 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe918f77-d000-44a8-9b16-c106f62cdede-config\") pod \"authentication-operator-69f744f599-pvqvn\" (UID: \"fe918f77-d000-44a8-9b16-c106f62cdede\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.234366 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.234432 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/221dbb7a-96df-41b4-9c28-240d63013e93-node-pullsecrets\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.234843 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/221dbb7a-96df-41b4-9c28-240d63013e93-audit-dir\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.235297 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.235351 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-audit-dir\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.235929 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/221dbb7a-96df-41b4-9c28-240d63013e93-etcd-serving-ca\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.236283 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/221dbb7a-96df-41b4-9c28-240d63013e93-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.236397 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2de90211-84d3-45fc-87eb-efb7d7f16a33-serving-cert\") pod \"openshift-config-operator-7777fb866f-vr48t\" (UID: \"2de90211-84d3-45fc-87eb-efb7d7f16a33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.236553 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-audit-policies\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.237279 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a743581d-083b-4400-a1f3-a73f1cfece31-serving-cert\") pod \"route-controller-manager-6576b87f9c-qc4x6\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.237433 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.237859 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abf290ff-6ae2-4354-a3df-7103c94e481e-config\") pod \"machine-approver-56656f9798-qcjvq\" (UID: \"abf290ff-6ae2-4354-a3df-7103c94e481e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.238844 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.238927 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/221dbb7a-96df-41b4-9c28-240d63013e93-etcd-client\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.239688 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-audit-policies\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.239738 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/221dbb7a-96df-41b4-9c28-240d63013e93-encryption-config\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.240527 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a743581d-083b-4400-a1f3-a73f1cfece31-client-ca\") pod \"route-controller-manager-6576b87f9c-qc4x6\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.241414 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-encryption-config\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.241792 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe918f77-d000-44a8-9b16-c106f62cdede-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pvqvn\" (UID: \"fe918f77-d000-44a8-9b16-c106f62cdede\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.241867 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe918f77-d000-44a8-9b16-c106f62cdede-service-ca-bundle\") pod \"authentication-operator-69f744f599-pvqvn\" (UID: \"fe918f77-d000-44a8-9b16-c106f62cdede\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.242508 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be098182-dd51-4a19-920a-a797a175124a-config\") pod \"machine-api-operator-5694c8668f-qw86w\" (UID: \"be098182-dd51-4a19-920a-a797a175124a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.242701 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/baaddd6b-cbef-451e-9315-2e64908dfaa2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-48kjd\" (UID: \"baaddd6b-cbef-451e-9315-2e64908dfaa2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.245166 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/abf290ff-6ae2-4354-a3df-7103c94e481e-machine-approver-tls\") pod \"machine-approver-56656f9798-qcjvq\" (UID: \"abf290ff-6ae2-4354-a3df-7103c94e481e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.245871 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/221dbb7a-96df-41b4-9c28-240d63013e93-serving-cert\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.246450 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.246822 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/221dbb7a-96df-41b4-9c28-240d63013e93-config\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.246990 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.247787 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/be098182-dd51-4a19-920a-a797a175124a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qw86w\" (UID: \"be098182-dd51-4a19-920a-a797a175124a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.248792 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe918f77-d000-44a8-9b16-c106f62cdede-serving-cert\") pod \"authentication-operator-69f744f599-pvqvn\" (UID: \"fe918f77-d000-44a8-9b16-c106f62cdede\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.249543 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/221dbb7a-96df-41b4-9c28-240d63013e93-audit\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.249840 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/abf290ff-6ae2-4354-a3df-7103c94e481e-auth-proxy-config\") pod \"machine-approver-56656f9798-qcjvq\" (UID: \"abf290ff-6ae2-4354-a3df-7103c94e481e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.250117 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.251103 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.251878 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.252904 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hftzc\" (UniqueName: \"kubernetes.io/projected/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-kube-api-access-hftzc\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.253300 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/baaddd6b-cbef-451e-9315-2e64908dfaa2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-48kjd\" (UID: \"baaddd6b-cbef-451e-9315-2e64908dfaa2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.252593 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/221dbb7a-96df-41b4-9c28-240d63013e93-image-import-ca\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.253458 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.253563 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2de90211-84d3-45fc-87eb-efb7d7f16a33-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vr48t\" (UID: \"2de90211-84d3-45fc-87eb-efb7d7f16a33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.254453 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.254611 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9prz\" (UniqueName: \"kubernetes.io/projected/fe918f77-d000-44a8-9b16-c106f62cdede-kube-api-access-w9prz\") pod \"authentication-operator-69f744f599-pvqvn\" (UID: \"fe918f77-d000-44a8-9b16-c106f62cdede\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.255986 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppv8z\" (UniqueName: \"kubernetes.io/projected/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-kube-api-access-ppv8z\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.257274 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b124c538-2e03-45ea-9666-b52b1fc90bdb-serving-cert\") pod \"controller-manager-879f6c89f-s4hdj\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.257801 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-serving-cert\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.258396 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-client-ca\") pod \"controller-manager-879f6c89f-s4hdj\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.264477 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.267136 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.267801 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/92e9d9e5-3307-46cb-80f8-0a944ba1bab4-etcd-client\") pod \"apiserver-7bbb656c7d-62tk9\" (UID: \"92e9d9e5-3307-46cb-80f8-0a944ba1bab4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.271758 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-789lm\" (UniqueName: \"kubernetes.io/projected/221dbb7a-96df-41b4-9c28-240d63013e93-kube-api-access-789lm\") pod \"apiserver-76f77b778f-hwgjb\" (UID: \"221dbb7a-96df-41b4-9c28-240d63013e93\") " pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.273133 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tjdp\" (UniqueName: \"kubernetes.io/projected/b124c538-2e03-45ea-9666-b52b1fc90bdb-kube-api-access-2tjdp\") pod \"controller-manager-879f6c89f-s4hdj\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.274978 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcv9g\" (UniqueName: \"kubernetes.io/projected/a743581d-083b-4400-a1f3-a73f1cfece31-kube-api-access-jcv9g\") pod \"route-controller-manager-6576b87f9c-qc4x6\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.275789 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.278408 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7kxt\" (UniqueName: \"kubernetes.io/projected/baaddd6b-cbef-451e-9315-2e64908dfaa2-kube-api-access-d7kxt\") pod \"openshift-apiserver-operator-796bbdcf4f-48kjd\" (UID: \"baaddd6b-cbef-451e-9315-2e64908dfaa2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.281700 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-grsv4\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.283397 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjhd6\" (UniqueName: \"kubernetes.io/projected/be098182-dd51-4a19-920a-a797a175124a-kube-api-access-hjhd6\") pod \"machine-api-operator-5694c8668f-qw86w\" (UID: \"be098182-dd51-4a19-920a-a797a175124a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.288033 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7q9v5"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.288777 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.307353 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhcwr\" (UniqueName: \"kubernetes.io/projected/abf290ff-6ae2-4354-a3df-7103c94e481e-kube-api-access-xhcwr\") pod \"machine-approver-56656f9798-qcjvq\" (UID: \"abf290ff-6ae2-4354-a3df-7103c94e481e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.314558 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.315578 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-262xt"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.316772 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.316779 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.317411 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.317880 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.318012 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.318280 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.318507 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.318766 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.319002 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.326606 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.326854 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.327498 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.328598 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.332875 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.333576 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.333805 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.336904 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.338068 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.338224 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sblhd\" (UniqueName: \"kubernetes.io/projected/caa5fa90-31b2-49f5-954e-9640fdd8140d-kube-api-access-sblhd\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.338333 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/caa5fa90-31b2-49f5-954e-9640fdd8140d-etcd-service-ca\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.338912 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.340207 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.340427 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.340624 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.341091 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.341431 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.342369 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzd2r\" (UniqueName: \"kubernetes.io/projected/2de90211-84d3-45fc-87eb-efb7d7f16a33-kube-api-access-vzd2r\") pod \"openshift-config-operator-7777fb866f-vr48t\" (UID: \"2de90211-84d3-45fc-87eb-efb7d7f16a33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.341561 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.341618 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.342420 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.366881 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.370604 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482c868b-ee7d-48d9-b9fb-8fb0527c95bf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pwq5g\" (UID: \"482c868b-ee7d-48d9-b9fb-8fb0527c95bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.370737 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caa5fa90-31b2-49f5-954e-9640fdd8140d-serving-cert\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.370763 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/caa5fa90-31b2-49f5-954e-9640fdd8140d-etcd-ca\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.370782 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/caa5fa90-31b2-49f5-954e-9640fdd8140d-etcd-client\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.370806 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/619b567a-92a9-4801-991e-792545b45bd0-serving-cert\") pod \"console-operator-58897d9998-262xt\" (UID: \"619b567a-92a9-4801-991e-792545b45bd0\") " pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.370827 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/482c868b-ee7d-48d9-b9fb-8fb0527c95bf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pwq5g\" (UID: \"482c868b-ee7d-48d9-b9fb-8fb0527c95bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.370860 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkq8l\" (UniqueName: \"kubernetes.io/projected/482c868b-ee7d-48d9-b9fb-8fb0527c95bf-kube-api-access-tkq8l\") pod \"openshift-controller-manager-operator-756b6f6bc6-pwq5g\" (UID: \"482c868b-ee7d-48d9-b9fb-8fb0527c95bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.370896 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caa5fa90-31b2-49f5-954e-9640fdd8140d-config\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.370917 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/619b567a-92a9-4801-991e-792545b45bd0-config\") pod \"console-operator-58897d9998-262xt\" (UID: \"619b567a-92a9-4801-991e-792545b45bd0\") " pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.370959 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhlqm\" (UniqueName: \"kubernetes.io/projected/619b567a-92a9-4801-991e-792545b45bd0-kube-api-access-dhlqm\") pod \"console-operator-58897d9998-262xt\" (UID: \"619b567a-92a9-4801-991e-792545b45bd0\") " pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.370996 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/619b567a-92a9-4801-991e-792545b45bd0-trusted-ca\") pod \"console-operator-58897d9998-262xt\" (UID: \"619b567a-92a9-4801-991e-792545b45bd0\") " pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.372488 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.375266 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.378810 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.378839 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pvqvn"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.378890 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.379877 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.386745 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-v8wh9"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.387574 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-v8wh9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.388172 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.388654 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-s4hdj"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.390552 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-c7pwq"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.391164 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.392373 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-nzh9r"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.393320 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-nzh9r" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.395400 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.395420 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fgjv4"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.396822 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fgjv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.396850 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-42p77"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.397574 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-n2cpp"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.398062 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n2cpp" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.398131 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-42p77" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.404048 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.404500 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.405228 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.405664 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.406172 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.406379 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-qnznl"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.406845 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-qnznl" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.407375 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vw9z9"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.407762 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vw9z9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.409448 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.409927 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.415392 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-x5gxr"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.416689 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.418243 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.418882 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.419367 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4hvgn"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.420549 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.424649 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.424847 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.425322 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.428591 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.431986 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.432452 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.432913 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.433217 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.433733 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.434527 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.435973 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.437628 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.438274 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.438488 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.439561 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.441227 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.442373 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.442713 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-bddbd"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.443243 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.443397 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.443775 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.444067 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.446993 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q58ld"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.447951 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-q58ld" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.449591 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.451464 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-w5q27"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.452152 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-w5q27" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.455927 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.466606 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-262xt"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.467222 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.469319 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.472573 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482c868b-ee7d-48d9-b9fb-8fb0527c95bf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pwq5g\" (UID: \"482c868b-ee7d-48d9-b9fb-8fb0527c95bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.472656 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caa5fa90-31b2-49f5-954e-9640fdd8140d-serving-cert\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.472692 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/caa5fa90-31b2-49f5-954e-9640fdd8140d-etcd-ca\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.472723 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/caa5fa90-31b2-49f5-954e-9640fdd8140d-etcd-client\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.472753 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/619b567a-92a9-4801-991e-792545b45bd0-serving-cert\") pod \"console-operator-58897d9998-262xt\" (UID: \"619b567a-92a9-4801-991e-792545b45bd0\") " pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.472785 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/482c868b-ee7d-48d9-b9fb-8fb0527c95bf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pwq5g\" (UID: \"482c868b-ee7d-48d9-b9fb-8fb0527c95bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.472830 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e8306af7-7828-4bf6-8099-3a986fd2572b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-w4hl6\" (UID: \"e8306af7-7828-4bf6-8099-3a986fd2572b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.472873 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkq8l\" (UniqueName: \"kubernetes.io/projected/482c868b-ee7d-48d9-b9fb-8fb0527c95bf-kube-api-access-tkq8l\") pod \"openshift-controller-manager-operator-756b6f6bc6-pwq5g\" (UID: \"482c868b-ee7d-48d9-b9fb-8fb0527c95bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.472908 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e8306af7-7828-4bf6-8099-3a986fd2572b-metrics-tls\") pod \"ingress-operator-5b745b69d9-w4hl6\" (UID: \"e8306af7-7828-4bf6-8099-3a986fd2572b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.474079 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caa5fa90-31b2-49f5-954e-9640fdd8140d-config\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.474139 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/619b567a-92a9-4801-991e-792545b45bd0-config\") pod \"console-operator-58897d9998-262xt\" (UID: \"619b567a-92a9-4801-991e-792545b45bd0\") " pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.474168 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhlqm\" (UniqueName: \"kubernetes.io/projected/619b567a-92a9-4801-991e-792545b45bd0-kube-api-access-dhlqm\") pod \"console-operator-58897d9998-262xt\" (UID: \"619b567a-92a9-4801-991e-792545b45bd0\") " pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.474206 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/619b567a-92a9-4801-991e-792545b45bd0-trusted-ca\") pod \"console-operator-58897d9998-262xt\" (UID: \"619b567a-92a9-4801-991e-792545b45bd0\") " pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.474262 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e8306af7-7828-4bf6-8099-3a986fd2572b-trusted-ca\") pod \"ingress-operator-5b745b69d9-w4hl6\" (UID: \"e8306af7-7828-4bf6-8099-3a986fd2572b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.474299 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jjzj\" (UniqueName: \"kubernetes.io/projected/e8306af7-7828-4bf6-8099-3a986fd2572b-kube-api-access-6jjzj\") pod \"ingress-operator-5b745b69d9-w4hl6\" (UID: \"e8306af7-7828-4bf6-8099-3a986fd2572b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.474330 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sblhd\" (UniqueName: \"kubernetes.io/projected/caa5fa90-31b2-49f5-954e-9640fdd8140d-kube-api-access-sblhd\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.474400 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/caa5fa90-31b2-49f5-954e-9640fdd8140d-etcd-service-ca\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.474591 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.478999 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/619b567a-92a9-4801-991e-792545b45bd0-serving-cert\") pod \"console-operator-58897d9998-262xt\" (UID: \"619b567a-92a9-4801-991e-792545b45bd0\") " pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.479997 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caa5fa90-31b2-49f5-954e-9640fdd8140d-config\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.480159 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/caa5fa90-31b2-49f5-954e-9640fdd8140d-etcd-ca\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.480402 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/482c868b-ee7d-48d9-b9fb-8fb0527c95bf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pwq5g\" (UID: \"482c868b-ee7d-48d9-b9fb-8fb0527c95bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.480494 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/619b567a-92a9-4801-991e-792545b45bd0-config\") pod \"console-operator-58897d9998-262xt\" (UID: \"619b567a-92a9-4801-991e-792545b45bd0\") " pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.482397 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/619b567a-92a9-4801-991e-792545b45bd0-trusted-ca\") pod \"console-operator-58897d9998-262xt\" (UID: \"619b567a-92a9-4801-991e-792545b45bd0\") " pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.483088 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.483714 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/caa5fa90-31b2-49f5-954e-9640fdd8140d-etcd-service-ca\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.483871 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482c868b-ee7d-48d9-b9fb-8fb0527c95bf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pwq5g\" (UID: \"482c868b-ee7d-48d9-b9fb-8fb0527c95bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.484330 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.487767 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.488698 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7q9v5"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.490443 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caa5fa90-31b2-49f5-954e-9640fdd8140d-serving-cert\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.493305 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.494376 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.496219 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/caa5fa90-31b2-49f5-954e-9640fdd8140d-etcd-client\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.497130 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.499242 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-g8kh5"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.504770 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.505688 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vw9z9"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.505871 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.508327 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fgjv4"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.512450 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-qnznl"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.514431 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-v8wh9"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.524671 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.539095 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.545474 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.545442 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-n2cpp"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.547239 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.549998 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-nzh9r"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.554026 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.554085 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.555382 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-c7pwq"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.564871 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.571792 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.577617 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e8306af7-7828-4bf6-8099-3a986fd2572b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-w4hl6\" (UID: \"e8306af7-7828-4bf6-8099-3a986fd2572b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.577692 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e8306af7-7828-4bf6-8099-3a986fd2572b-metrics-tls\") pod \"ingress-operator-5b745b69d9-w4hl6\" (UID: \"e8306af7-7828-4bf6-8099-3a986fd2572b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.577762 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e8306af7-7828-4bf6-8099-3a986fd2572b-trusted-ca\") pod \"ingress-operator-5b745b69d9-w4hl6\" (UID: \"e8306af7-7828-4bf6-8099-3a986fd2572b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.577787 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jjzj\" (UniqueName: \"kubernetes.io/projected/e8306af7-7828-4bf6-8099-3a986fd2572b-kube-api-access-6jjzj\") pod \"ingress-operator-5b745b69d9-w4hl6\" (UID: \"e8306af7-7828-4bf6-8099-3a986fd2572b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.582888 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.586871 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.592732 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.595296 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.599575 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.599604 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4hvgn"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.601448 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.604825 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.617366 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.619655 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-g8kh5"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.619711 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-42p77"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.623077 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.624593 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-x5gxr"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.625663 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.627145 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-v4fcn"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.628052 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q58ld"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.628160 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-v4fcn" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.629667 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.629816 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-v4fcn"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.646185 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.669157 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.683832 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.706220 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.723782 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.768036 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.768117 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.789353 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.811386 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.824050 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.848834 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.869667 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.870285 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.883988 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.890900 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hwgjb"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.904848 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.905430 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.905468 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.906887 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pvqvn"] Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.923161 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.965922 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.967061 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.985458 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 17:27:55 crc kubenswrapper[4868]: I1201 17:27:55.996923 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qw86w"] Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.006585 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.009795 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"95bbdbdfd8a6338908f805558afd044b5609c77dbe46069c0b6158a9f7111285"} Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.016179 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" event={"ID":"fe918f77-d000-44a8-9b16-c106f62cdede","Type":"ContainerStarted","Data":"821caf79bbd4ae94e84fb356aea3010f79b0d2d377a7da69cb85a7bcbf82ada5"} Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.019403 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" event={"ID":"a743581d-083b-4400-a1f3-a73f1cfece31","Type":"ContainerStarted","Data":"8d44d29517630feb4ed140233c1e936f90449d7f512e0c9d9831169ec1a9a8ad"} Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.020366 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" event={"ID":"abf290ff-6ae2-4354-a3df-7103c94e481e","Type":"ContainerStarted","Data":"efa8a759418c4c2735e83d2bce5a3c52fbbe95ce0d0b7db6ff8d2f6f0da4c470"} Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.024579 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.024704 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" event={"ID":"221dbb7a-96df-41b4-9c28-240d63013e93","Type":"ContainerStarted","Data":"e2d9ea31255ea7c039aff2b294bd574a343489dc9600501fd61ce214886b752f"} Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.033589 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vr48t"] Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.044764 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 17:27:56 crc kubenswrapper[4868]: W1201 17:27:56.059336 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2de90211_84d3_45fc_87eb_efb7d7f16a33.slice/crio-78f7a502e6a8d661c66750432b32d2ba9f9e84f6593c0495bf7f25a6733ef97d WatchSource:0}: Error finding container 78f7a502e6a8d661c66750432b32d2ba9f9e84f6593c0495bf7f25a6733ef97d: Status 404 returned error can't find the container with id 78f7a502e6a8d661c66750432b32d2ba9f9e84f6593c0495bf7f25a6733ef97d Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.063834 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.088404 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.107011 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.123591 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9"] Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.124382 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.130876 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd"] Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.147705 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.163282 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 17:27:56 crc kubenswrapper[4868]: W1201 17:27:56.165388 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-e6248232f8a6a20e9bb7b655a4219bf89157e0f1e43466527fba53bde5310b38 WatchSource:0}: Error finding container e6248232f8a6a20e9bb7b655a4219bf89157e0f1e43466527fba53bde5310b38: Status 404 returned error can't find the container with id e6248232f8a6a20e9bb7b655a4219bf89157e0f1e43466527fba53bde5310b38 Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.187086 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.204451 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.224696 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.262904 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.274427 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e8306af7-7828-4bf6-8099-3a986fd2572b-trusted-ca\") pod \"ingress-operator-5b745b69d9-w4hl6\" (UID: \"e8306af7-7828-4bf6-8099-3a986fd2572b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.280040 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.289573 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.322904 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.325001 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-s4hdj"] Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.327022 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.349338 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e8306af7-7828-4bf6-8099-3a986fd2572b-metrics-tls\") pod \"ingress-operator-5b745b69d9-w4hl6\" (UID: \"e8306af7-7828-4bf6-8099-3a986fd2572b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.350412 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.352687 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-grsv4"] Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.369100 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.404551 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.423682 4868 request.go:700] Waited for 1.002791953s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/secrets?fieldSelector=metadata.name%3Dmarketplace-operator-dockercfg-5nsgg&limit=500&resourceVersion=0 Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.426547 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.452183 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.465350 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.483614 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.504553 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.526469 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.543323 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.564778 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.584560 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.603823 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.624377 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.644016 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.665551 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.685231 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.704436 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.724031 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.743682 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.764270 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.783896 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.804161 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.842997 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.864749 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.891535 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.906108 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.923757 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.945455 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.963508 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 17:27:56 crc kubenswrapper[4868]: I1201 17:27:56.984925 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.004454 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.024442 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.044353 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.051647 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" event={"ID":"be098182-dd51-4a19-920a-a797a175124a","Type":"ContainerStarted","Data":"f8b6a59304c2170e8ccdde9c020892d17123469f19e33a4c2e71db09658c1b5f"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.051710 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" event={"ID":"be098182-dd51-4a19-920a-a797a175124a","Type":"ContainerStarted","Data":"3683270c4de69539e2d0dca0f82f2f1ed2367c97577399178767f3bd446ebd07"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.051728 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" event={"ID":"be098182-dd51-4a19-920a-a797a175124a","Type":"ContainerStarted","Data":"961be66ab1f92b35f749cf0832862949998cc8e7671218d8122f5e356d5846d3"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.054504 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"cff13e750bc69ae7bc210295b793eabedd9f8a2e0e8e29d758237ca8a4ca21eb"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.054541 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e6248232f8a6a20e9bb7b655a4219bf89157e0f1e43466527fba53bde5310b38"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.054712 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.055749 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" event={"ID":"a743581d-083b-4400-a1f3-a73f1cfece31","Type":"ContainerStarted","Data":"3c5284e965f3be4e86608e4bd1301c2443e174e3ecc7737c958454d7d9c9675c"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.055971 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.057608 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" event={"ID":"abf290ff-6ae2-4354-a3df-7103c94e481e","Type":"ContainerStarted","Data":"2a05d8fed8a3fbd44c906ad55f26f2c155d7e83db631e6c6f9164f1dfb4c94ed"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.057644 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" event={"ID":"abf290ff-6ae2-4354-a3df-7103c94e481e","Type":"ContainerStarted","Data":"ec2a5b1d297670d750ad923892972b60dde5397e6b7d60b80415d1a06f2dd1ec"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.061562 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd" event={"ID":"baaddd6b-cbef-451e-9315-2e64908dfaa2","Type":"ContainerStarted","Data":"4728730a5d01dbc610beb59fedb65de9bab0f37421f1ec8b8aad88ba5e8d7eef"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.061653 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd" event={"ID":"baaddd6b-cbef-451e-9315-2e64908dfaa2","Type":"ContainerStarted","Data":"f5d36897b99864bacf40be2999369afd34fd34b12be231f21c2a92aec2e59779"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.063958 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"fe9c1cf0f342ff60bcb043638d687f7b3c06e8bfc4bab10790f19c20f5342cf9"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.066228 4868 generic.go:334] "Generic (PLEG): container finished" podID="92e9d9e5-3307-46cb-80f8-0a944ba1bab4" containerID="82c9e9598a580828d68c6ec0847986125f54946e346815e2cd82b837fa35fb2a" exitCode=0 Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.066297 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" event={"ID":"92e9d9e5-3307-46cb-80f8-0a944ba1bab4","Type":"ContainerDied","Data":"82c9e9598a580828d68c6ec0847986125f54946e346815e2cd82b837fa35fb2a"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.066325 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" event={"ID":"92e9d9e5-3307-46cb-80f8-0a944ba1bab4","Type":"ContainerStarted","Data":"ac2ff72c262cc67eb5699f1a00c205769b76045f404b2a8792dfe70585770c07"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.070126 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" event={"ID":"fe918f77-d000-44a8-9b16-c106f62cdede","Type":"ContainerStarted","Data":"43b4b22bfe3cc0339b856ab49ab199b4f5ac071909f0ec68dfd7bdde1c594058"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.072581 4868 generic.go:334] "Generic (PLEG): container finished" podID="2de90211-84d3-45fc-87eb-efb7d7f16a33" containerID="fcde73fa4f5d0dd36e0d903a310ebfc4191570fcad8d1cd3be4ee01d52a3117f" exitCode=0 Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.072646 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" event={"ID":"2de90211-84d3-45fc-87eb-efb7d7f16a33","Type":"ContainerDied","Data":"fcde73fa4f5d0dd36e0d903a310ebfc4191570fcad8d1cd3be4ee01d52a3117f"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.072672 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" event={"ID":"2de90211-84d3-45fc-87eb-efb7d7f16a33","Type":"ContainerStarted","Data":"78f7a502e6a8d661c66750432b32d2ba9f9e84f6593c0495bf7f25a6733ef97d"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.075232 4868 generic.go:334] "Generic (PLEG): container finished" podID="221dbb7a-96df-41b4-9c28-240d63013e93" containerID="601a72287743dccaa3ba3506b6a4c14dc55980cccc91a1690bbb2db8d75b996a" exitCode=0 Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.075286 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" event={"ID":"221dbb7a-96df-41b4-9c28-240d63013e93","Type":"ContainerDied","Data":"601a72287743dccaa3ba3506b6a4c14dc55980cccc91a1690bbb2db8d75b996a"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.092089 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7719ff37a261b30164742e8c54c3766411b713121be4fcfd8639a3cc08642a2b"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.092142 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"20807e48fedf917e2640fedda932f4326ac5afa0622e69948a1037df781f371c"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.094472 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" event={"ID":"6749e2f1-9c2c-41ce-8487-44f1a7ff7576","Type":"ContainerStarted","Data":"5599db4f6b49491c9ecd619e10ef3425d7732c6c60af4c98188127a421e6ddcf"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.094528 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" event={"ID":"6749e2f1-9c2c-41ce-8487-44f1a7ff7576","Type":"ContainerStarted","Data":"4de7f02a896425ddc2f2fd860b389e7f856e98f308a8766b06565a35ceedcabc"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.094690 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.097186 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" event={"ID":"b124c538-2e03-45ea-9666-b52b1fc90bdb","Type":"ContainerStarted","Data":"902a78ae3ce382dd36022c1cbe55a8af2c736995e0d92b8f0fe25eec50dd7da5"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.097230 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" event={"ID":"b124c538-2e03-45ea-9666-b52b1fc90bdb","Type":"ContainerStarted","Data":"21908cffef69ec16cb38ee5e8c1c45e04afbad581d2998ac1906b68df264afce"} Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.097892 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.099994 4868 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-grsv4 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.100082 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" podUID="6749e2f1-9c2c-41ce-8487-44f1a7ff7576" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.101823 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.102124 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.103717 4868 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-s4hdj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.103773 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" podUID="b124c538-2e03-45ea-9666-b52b1fc90bdb" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.104564 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.123521 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.144208 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.166012 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.184385 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.203934 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.224592 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.236367 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.244161 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.323818 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkq8l\" (UniqueName: \"kubernetes.io/projected/482c868b-ee7d-48d9-b9fb-8fb0527c95bf-kube-api-access-tkq8l\") pod \"openshift-controller-manager-operator-756b6f6bc6-pwq5g\" (UID: \"482c868b-ee7d-48d9-b9fb-8fb0527c95bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.323849 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sblhd\" (UniqueName: \"kubernetes.io/projected/caa5fa90-31b2-49f5-954e-9640fdd8140d-kube-api-access-sblhd\") pod \"etcd-operator-b45778765-7q9v5\" (UID: \"caa5fa90-31b2-49f5-954e-9640fdd8140d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.325512 4868 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.332028 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhlqm\" (UniqueName: \"kubernetes.io/projected/619b567a-92a9-4801-991e-792545b45bd0-kube-api-access-dhlqm\") pod \"console-operator-58897d9998-262xt\" (UID: \"619b567a-92a9-4801-991e-792545b45bd0\") " pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.346879 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.380663 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.405560 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jjzj\" (UniqueName: \"kubernetes.io/projected/e8306af7-7828-4bf6-8099-3a986fd2572b-kube-api-access-6jjzj\") pod \"ingress-operator-5b745b69d9-w4hl6\" (UID: \"e8306af7-7828-4bf6-8099-3a986fd2572b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.424588 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.441507 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e8306af7-7828-4bf6-8099-3a986fd2572b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-w4hl6\" (UID: \"e8306af7-7828-4bf6-8099-3a986fd2572b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.441843 4868 request.go:700] Waited for 1.813382248s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Ddefault-dockercfg-2llfx&limit=500&resourceVersion=0 Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.443634 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.464103 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.483454 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.484462 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.484961 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.517307 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.525201 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.534000 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-bound-sa-token\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.534070 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bpgd\" (UniqueName: \"kubernetes.io/projected/dcd0102b-419f-4d81-8187-af734155ca55-kube-api-access-9bpgd\") pod \"machine-config-operator-74547568cd-59gxf\" (UID: \"dcd0102b-419f-4d81-8187-af734155ca55\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.534103 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74782477-7d4a-498c-a3dd-339cebc5ae37-config\") pod \"service-ca-operator-777779d784-v4hbg\" (UID: \"74782477-7d4a-498c-a3dd-339cebc5ae37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.534129 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctll7\" (UniqueName: \"kubernetes.io/projected/74782477-7d4a-498c-a3dd-339cebc5ae37-kube-api-access-ctll7\") pod \"service-ca-operator-777779d784-v4hbg\" (UID: \"74782477-7d4a-498c-a3dd-339cebc5ae37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.534152 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f66e6c7-0b8b-4ef9-b848-ce56e2501193-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljt2x\" (UID: \"1f66e6c7-0b8b-4ef9-b848-ce56e2501193\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.534173 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6-proxy-tls\") pod \"machine-config-controller-84d6567774-k7ghg\" (UID: \"e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.534197 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39ac4687-0a95-43b1-a4b1-334ae97544d2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gkxh9\" (UID: \"39ac4687-0a95-43b1-a4b1-334ae97544d2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.534241 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.534282 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-service-ca\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.534320 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-k7ghg\" (UID: \"e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.535582 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f66e6c7-0b8b-4ef9-b848-ce56e2501193-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljt2x\" (UID: \"1f66e6c7-0b8b-4ef9-b848-ce56e2501193\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.535642 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9klcn\" (UniqueName: \"kubernetes.io/projected/9307075d-5039-4f41-ac9a-8d6c2dbca531-kube-api-access-9klcn\") pod \"downloads-7954f5f757-qnznl\" (UID: \"9307075d-5039-4f41-ac9a-8d6c2dbca531\") " pod="openshift-console/downloads-7954f5f757-qnznl" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.535679 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-trusted-ca\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.535702 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dcd0102b-419f-4d81-8187-af734155ca55-images\") pod \"machine-config-operator-74547568cd-59gxf\" (UID: \"dcd0102b-419f-4d81-8187-af734155ca55\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.535789 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dcd0102b-419f-4d81-8187-af734155ca55-auth-proxy-config\") pod \"machine-config-operator-74547568cd-59gxf\" (UID: \"dcd0102b-419f-4d81-8187-af734155ca55\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.536605 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-console-config\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.538072 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/313d7661-5eec-4a28-92a7-cf1f4f85d2e6-config\") pod \"kube-controller-manager-operator-78b949d7b-c4nzd\" (UID: \"313d7661-5eec-4a28-92a7-cf1f4f85d2e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.538447 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9a983c06-7704-4c22-8c1b-d0cfcdc67ee8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fgjv4\" (UID: \"9a983c06-7704-4c22-8c1b-d0cfcdc67ee8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fgjv4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.539124 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk4t2\" (UniqueName: \"kubernetes.io/projected/8cddeb4c-690f-4627-b04a-fedf2fce8594-kube-api-access-nk4t2\") pod \"migrator-59844c95c7-n2cpp\" (UID: \"8cddeb4c-690f-4627-b04a-fedf2fce8594\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n2cpp" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.540375 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nwzv\" (UniqueName: \"kubernetes.io/projected/e90813d0-8960-450f-aaa4-047341c50947-kube-api-access-4nwzv\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.542104 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcjqp\" (UniqueName: \"kubernetes.io/projected/39ac4687-0a95-43b1-a4b1-334ae97544d2-kube-api-access-dcjqp\") pod \"kube-storage-version-migrator-operator-b67b599dd-gkxh9\" (UID: \"39ac4687-0a95-43b1-a4b1-334ae97544d2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.542642 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-registry-tls\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.543091 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkkpz\" (UniqueName: \"kubernetes.io/projected/7ca45004-4f76-480b-b8b8-68178408dde2-kube-api-access-bkkpz\") pod \"cluster-samples-operator-665b6dd947-42p77\" (UID: \"7ca45004-4f76-480b-b8b8-68178408dde2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-42p77" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.543342 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-oauth-serving-cert\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.544018 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-trusted-ca-bundle\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.545997 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39ac4687-0a95-43b1-a4b1-334ae97544d2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gkxh9\" (UID: \"39ac4687-0a95-43b1-a4b1-334ae97544d2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.546036 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/313d7661-5eec-4a28-92a7-cf1f4f85d2e6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-c4nzd\" (UID: \"313d7661-5eec-4a28-92a7-cf1f4f85d2e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.546057 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jflll\" (UniqueName: \"kubernetes.io/projected/9a983c06-7704-4c22-8c1b-d0cfcdc67ee8-kube-api-access-jflll\") pod \"multus-admission-controller-857f4d67dd-fgjv4\" (UID: \"9a983c06-7704-4c22-8c1b-d0cfcdc67ee8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fgjv4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.546092 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-registry-certificates\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.550603 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74782477-7d4a-498c-a3dd-339cebc5ae37-serving-cert\") pod \"service-ca-operator-777779d784-v4hbg\" (UID: \"74782477-7d4a-498c-a3dd-339cebc5ae37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.550834 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/7ca45004-4f76-480b-b8b8-68178408dde2-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-42p77\" (UID: \"7ca45004-4f76-480b-b8b8-68178408dde2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-42p77" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.551203 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e90813d0-8960-450f-aaa4-047341c50947-console-oauth-config\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.551574 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: E1201 17:27:57.553717 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:27:58.05369273 +0000 UTC m=+150.424803141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.551739 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/313d7661-5eec-4a28-92a7-cf1f4f85d2e6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-c4nzd\" (UID: \"313d7661-5eec-4a28-92a7-cf1f4f85d2e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.555639 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e90813d0-8960-450f-aaa4-047341c50947-console-serving-cert\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.555728 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4d64\" (UniqueName: \"kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-kube-api-access-x4d64\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.555772 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dcd0102b-419f-4d81-8187-af734155ca55-proxy-tls\") pod \"machine-config-operator-74547568cd-59gxf\" (UID: \"dcd0102b-419f-4d81-8187-af734155ca55\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.557963 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f66e6c7-0b8b-4ef9-b848-ce56e2501193-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljt2x\" (UID: \"1f66e6c7-0b8b-4ef9-b848-ce56e2501193\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.558050 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.558081 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xdwt\" (UniqueName: \"kubernetes.io/projected/3bfca371-4e3e-4373-9589-1e1ee7042f85-kube-api-access-8xdwt\") pod \"dns-operator-744455d44c-v8wh9\" (UID: \"3bfca371-4e3e-4373-9589-1e1ee7042f85\") " pod="openshift-dns-operator/dns-operator-744455d44c-v8wh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.558112 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3bfca371-4e3e-4373-9589-1e1ee7042f85-metrics-tls\") pod \"dns-operator-744455d44c-v8wh9\" (UID: \"3bfca371-4e3e-4373-9589-1e1ee7042f85\") " pod="openshift-dns-operator/dns-operator-744455d44c-v8wh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.558186 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-828lj\" (UniqueName: \"kubernetes.io/projected/e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6-kube-api-access-828lj\") pod \"machine-config-controller-84d6567774-k7ghg\" (UID: \"e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.665405 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.665655 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nqj9\" (UniqueName: \"kubernetes.io/projected/22cc1879-5e5e-4a06-ba88-f92033382b90-kube-api-access-9nqj9\") pod \"marketplace-operator-79b997595-4hvgn\" (UID: \"22cc1879-5e5e-4a06-ba88-f92033382b90\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.665706 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/27ae76b5-fd48-4e32-9bf3-a8813f08762b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5dqtb\" (UID: \"27ae76b5-fd48-4e32-9bf3-a8813f08762b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.665726 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a5bca2c5-996d-49a5-92ac-1d38c96134eb-signing-key\") pod \"service-ca-9c57cc56f-q58ld\" (UID: \"a5bca2c5-996d-49a5-92ac-1d38c96134eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-q58ld" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.665771 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzc55\" (UniqueName: \"kubernetes.io/projected/886b276a-b6e7-4ce1-b6cf-dc269d9180bb-kube-api-access-tzc55\") pod \"router-default-5444994796-bddbd\" (UID: \"886b276a-b6e7-4ce1-b6cf-dc269d9180bb\") " pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.665790 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rsjs\" (UniqueName: \"kubernetes.io/projected/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-kube-api-access-2rsjs\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.665814 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nwzv\" (UniqueName: \"kubernetes.io/projected/e90813d0-8960-450f-aaa4-047341c50947-kube-api-access-4nwzv\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.665831 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcjqp\" (UniqueName: \"kubernetes.io/projected/39ac4687-0a95-43b1-a4b1-334ae97544d2-kube-api-access-dcjqp\") pod \"kube-storage-version-migrator-operator-b67b599dd-gkxh9\" (UID: \"39ac4687-0a95-43b1-a4b1-334ae97544d2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.665849 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/da5b10e6-e695-4e8a-84e5-329870d5736e-tmpfs\") pod \"packageserver-d55dfcdfc-wnnm4\" (UID: \"da5b10e6-e695-4e8a-84e5-329870d5736e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.665866 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k498r\" (UniqueName: \"kubernetes.io/projected/7b50f7b4-dd55-4255-9b68-a942ff4a086e-kube-api-access-k498r\") pod \"dns-default-nzh9r\" (UID: \"7b50f7b4-dd55-4255-9b68-a942ff4a086e\") " pod="openshift-dns/dns-default-nzh9r" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.665884 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m648p\" (UniqueName: \"kubernetes.io/projected/5e59c27e-823b-4ecf-9ccf-c198528219e9-kube-api-access-m648p\") pod \"cluster-image-registry-operator-dc59b4c8b-m8zhv\" (UID: \"5e59c27e-823b-4ecf-9ccf-c198528219e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.665904 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-registry-tls\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.665919 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/886b276a-b6e7-4ce1-b6cf-dc269d9180bb-default-certificate\") pod \"router-default-5444994796-bddbd\" (UID: \"886b276a-b6e7-4ce1-b6cf-dc269d9180bb\") " pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.665956 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02-certs\") pod \"machine-config-server-w5q27\" (UID: \"b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02\") " pod="openshift-machine-config-operator/machine-config-server-w5q27" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.665980 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a5bca2c5-996d-49a5-92ac-1d38c96134eb-signing-cabundle\") pod \"service-ca-9c57cc56f-q58ld\" (UID: \"a5bca2c5-996d-49a5-92ac-1d38c96134eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-q58ld" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666005 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkkpz\" (UniqueName: \"kubernetes.io/projected/7ca45004-4f76-480b-b8b8-68178408dde2-kube-api-access-bkkpz\") pod \"cluster-samples-operator-665b6dd947-42p77\" (UID: \"7ca45004-4f76-480b-b8b8-68178408dde2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-42p77" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666026 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02-node-bootstrap-token\") pod \"machine-config-server-w5q27\" (UID: \"b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02\") " pod="openshift-machine-config-operator/machine-config-server-w5q27" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666045 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-oauth-serving-cert\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666061 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn8pw\" (UniqueName: \"kubernetes.io/projected/da5b10e6-e695-4e8a-84e5-329870d5736e-kube-api-access-kn8pw\") pod \"packageserver-d55dfcdfc-wnnm4\" (UID: \"da5b10e6-e695-4e8a-84e5-329870d5736e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666077 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5e59c27e-823b-4ecf-9ccf-c198528219e9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-m8zhv\" (UID: \"5e59c27e-823b-4ecf-9ccf-c198528219e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666100 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-trusted-ca-bundle\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666123 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/962e4f14-b280-4ffa-bde0-6cc4b0d86144-secret-volume\") pod \"collect-profiles-29410155-bvw5m\" (UID: \"962e4f14-b280-4ffa-bde0-6cc4b0d86144\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666139 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-csi-data-dir\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666157 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/95e7a05d-5fa9-47ce-8cf3-d261fea05465-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4njnk\" (UID: \"95e7a05d-5fa9-47ce-8cf3-d261fea05465\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666176 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b0f5f4f9-02b9-48f1-918e-985b836626ad-cert\") pod \"ingress-canary-v4fcn\" (UID: \"b0f5f4f9-02b9-48f1-918e-985b836626ad\") " pod="openshift-ingress-canary/ingress-canary-v4fcn" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666195 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39ac4687-0a95-43b1-a4b1-334ae97544d2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gkxh9\" (UID: \"39ac4687-0a95-43b1-a4b1-334ae97544d2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666215 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74782477-7d4a-498c-a3dd-339cebc5ae37-serving-cert\") pod \"service-ca-operator-777779d784-v4hbg\" (UID: \"74782477-7d4a-498c-a3dd-339cebc5ae37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666236 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/313d7661-5eec-4a28-92a7-cf1f4f85d2e6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-c4nzd\" (UID: \"313d7661-5eec-4a28-92a7-cf1f4f85d2e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666259 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jflll\" (UniqueName: \"kubernetes.io/projected/9a983c06-7704-4c22-8c1b-d0cfcdc67ee8-kube-api-access-jflll\") pod \"multus-admission-controller-857f4d67dd-fgjv4\" (UID: \"9a983c06-7704-4c22-8c1b-d0cfcdc67ee8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fgjv4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666286 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-registry-certificates\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666303 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/7ca45004-4f76-480b-b8b8-68178408dde2-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-42p77\" (UID: \"7ca45004-4f76-480b-b8b8-68178408dde2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-42p77" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666319 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-plugins-dir\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666367 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f54tb\" (UniqueName: \"kubernetes.io/projected/95e7a05d-5fa9-47ce-8cf3-d261fea05465-kube-api-access-f54tb\") pod \"package-server-manager-789f6589d5-4njnk\" (UID: \"95e7a05d-5fa9-47ce-8cf3-d261fea05465\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666395 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e90813d0-8960-450f-aaa4-047341c50947-console-oauth-config\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666421 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27ae76b5-fd48-4e32-9bf3-a8813f08762b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5dqtb\" (UID: \"27ae76b5-fd48-4e32-9bf3-a8813f08762b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666450 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/da5b10e6-e695-4e8a-84e5-329870d5736e-webhook-cert\") pod \"packageserver-d55dfcdfc-wnnm4\" (UID: \"da5b10e6-e695-4e8a-84e5-329870d5736e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666468 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e3a5f31b-ac9f-4828-82bf-80100560da0c-srv-cert\") pod \"catalog-operator-68c6474976-mt2zl\" (UID: \"e3a5f31b-ac9f-4828-82bf-80100560da0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666486 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvc2p\" (UniqueName: \"kubernetes.io/projected/b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02-kube-api-access-fvc2p\") pod \"machine-config-server-w5q27\" (UID: \"b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02\") " pod="openshift-machine-config-operator/machine-config-server-w5q27" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666505 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/313d7661-5eec-4a28-92a7-cf1f4f85d2e6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-c4nzd\" (UID: \"313d7661-5eec-4a28-92a7-cf1f4f85d2e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666522 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4d64\" (UniqueName: \"kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-kube-api-access-x4d64\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666538 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e90813d0-8960-450f-aaa4-047341c50947-console-serving-cert\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666556 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5e59c27e-823b-4ecf-9ccf-c198528219e9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-m8zhv\" (UID: \"5e59c27e-823b-4ecf-9ccf-c198528219e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666576 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dcd0102b-419f-4d81-8187-af734155ca55-proxy-tls\") pod \"machine-config-operator-74547568cd-59gxf\" (UID: \"dcd0102b-419f-4d81-8187-af734155ca55\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666592 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e59c27e-823b-4ecf-9ccf-c198528219e9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-m8zhv\" (UID: \"5e59c27e-823b-4ecf-9ccf-c198528219e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666609 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22cc1879-5e5e-4a06-ba88-f92033382b90-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4hvgn\" (UID: \"22cc1879-5e5e-4a06-ba88-f92033382b90\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666632 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/962e4f14-b280-4ffa-bde0-6cc4b0d86144-config-volume\") pod \"collect-profiles-29410155-bvw5m\" (UID: \"962e4f14-b280-4ffa-bde0-6cc4b0d86144\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666652 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f66e6c7-0b8b-4ef9-b848-ce56e2501193-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljt2x\" (UID: \"1f66e6c7-0b8b-4ef9-b848-ce56e2501193\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666671 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-registration-dir\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666692 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666709 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xdwt\" (UniqueName: \"kubernetes.io/projected/3bfca371-4e3e-4373-9589-1e1ee7042f85-kube-api-access-8xdwt\") pod \"dns-operator-744455d44c-v8wh9\" (UID: \"3bfca371-4e3e-4373-9589-1e1ee7042f85\") " pod="openshift-dns-operator/dns-operator-744455d44c-v8wh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666729 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3bfca371-4e3e-4373-9589-1e1ee7042f85-metrics-tls\") pod \"dns-operator-744455d44c-v8wh9\" (UID: \"3bfca371-4e3e-4373-9589-1e1ee7042f85\") " pod="openshift-dns-operator/dns-operator-744455d44c-v8wh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666749 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7b50f7b4-dd55-4255-9b68-a942ff4a086e-metrics-tls\") pod \"dns-default-nzh9r\" (UID: \"7b50f7b4-dd55-4255-9b68-a942ff4a086e\") " pod="openshift-dns/dns-default-nzh9r" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666772 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e3a5f31b-ac9f-4828-82bf-80100560da0c-profile-collector-cert\") pod \"catalog-operator-68c6474976-mt2zl\" (UID: \"e3a5f31b-ac9f-4828-82bf-80100560da0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666793 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-828lj\" (UniqueName: \"kubernetes.io/projected/e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6-kube-api-access-828lj\") pod \"machine-config-controller-84d6567774-k7ghg\" (UID: \"e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666814 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/71ff3201-293e-4a9e-ae2a-006c945714ef-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7q4rx\" (UID: \"71ff3201-293e-4a9e-ae2a-006c945714ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666832 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z625w\" (UniqueName: \"kubernetes.io/projected/71ff3201-293e-4a9e-ae2a-006c945714ef-kube-api-access-z625w\") pod \"olm-operator-6b444d44fb-7q4rx\" (UID: \"71ff3201-293e-4a9e-ae2a-006c945714ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666849 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-bound-sa-token\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666865 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bpgd\" (UniqueName: \"kubernetes.io/projected/dcd0102b-419f-4d81-8187-af734155ca55-kube-api-access-9bpgd\") pod \"machine-config-operator-74547568cd-59gxf\" (UID: \"dcd0102b-419f-4d81-8187-af734155ca55\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666899 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74782477-7d4a-498c-a3dd-339cebc5ae37-config\") pod \"service-ca-operator-777779d784-v4hbg\" (UID: \"74782477-7d4a-498c-a3dd-339cebc5ae37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666917 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctll7\" (UniqueName: \"kubernetes.io/projected/74782477-7d4a-498c-a3dd-339cebc5ae37-kube-api-access-ctll7\") pod \"service-ca-operator-777779d784-v4hbg\" (UID: \"74782477-7d4a-498c-a3dd-339cebc5ae37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666933 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f66e6c7-0b8b-4ef9-b848-ce56e2501193-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljt2x\" (UID: \"1f66e6c7-0b8b-4ef9-b848-ce56e2501193\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666975 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv7k4\" (UniqueName: \"kubernetes.io/projected/e3a5f31b-ac9f-4828-82bf-80100560da0c-kube-api-access-gv7k4\") pod \"catalog-operator-68c6474976-mt2zl\" (UID: \"e3a5f31b-ac9f-4828-82bf-80100560da0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.666994 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm6hg\" (UniqueName: \"kubernetes.io/projected/962e4f14-b280-4ffa-bde0-6cc4b0d86144-kube-api-access-nm6hg\") pod \"collect-profiles-29410155-bvw5m\" (UID: \"962e4f14-b280-4ffa-bde0-6cc4b0d86144\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667012 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/886b276a-b6e7-4ce1-b6cf-dc269d9180bb-stats-auth\") pod \"router-default-5444994796-bddbd\" (UID: \"886b276a-b6e7-4ce1-b6cf-dc269d9180bb\") " pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667031 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39ac4687-0a95-43b1-a4b1-334ae97544d2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gkxh9\" (UID: \"39ac4687-0a95-43b1-a4b1-334ae97544d2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667047 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6-proxy-tls\") pod \"machine-config-controller-84d6567774-k7ghg\" (UID: \"e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667064 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667105 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-service-ca\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667122 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-socket-dir\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667148 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/886b276a-b6e7-4ce1-b6cf-dc269d9180bb-service-ca-bundle\") pod \"router-default-5444994796-bddbd\" (UID: \"886b276a-b6e7-4ce1-b6cf-dc269d9180bb\") " pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667167 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-k7ghg\" (UID: \"e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667184 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/22cc1879-5e5e-4a06-ba88-f92033382b90-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4hvgn\" (UID: \"22cc1879-5e5e-4a06-ba88-f92033382b90\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667230 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/71ff3201-293e-4a9e-ae2a-006c945714ef-srv-cert\") pod \"olm-operator-6b444d44fb-7q4rx\" (UID: \"71ff3201-293e-4a9e-ae2a-006c945714ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667266 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/886b276a-b6e7-4ce1-b6cf-dc269d9180bb-metrics-certs\") pod \"router-default-5444994796-bddbd\" (UID: \"886b276a-b6e7-4ce1-b6cf-dc269d9180bb\") " pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667281 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbf6r\" (UniqueName: \"kubernetes.io/projected/a5bca2c5-996d-49a5-92ac-1d38c96134eb-kube-api-access-zbf6r\") pod \"service-ca-9c57cc56f-q58ld\" (UID: \"a5bca2c5-996d-49a5-92ac-1d38c96134eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-q58ld" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667300 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dcd0102b-419f-4d81-8187-af734155ca55-images\") pod \"machine-config-operator-74547568cd-59gxf\" (UID: \"dcd0102b-419f-4d81-8187-af734155ca55\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667321 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f66e6c7-0b8b-4ef9-b848-ce56e2501193-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljt2x\" (UID: \"1f66e6c7-0b8b-4ef9-b848-ce56e2501193\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667339 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9klcn\" (UniqueName: \"kubernetes.io/projected/9307075d-5039-4f41-ac9a-8d6c2dbca531-kube-api-access-9klcn\") pod \"downloads-7954f5f757-qnznl\" (UID: \"9307075d-5039-4f41-ac9a-8d6c2dbca531\") " pod="openshift-console/downloads-7954f5f757-qnznl" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667357 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27ae76b5-fd48-4e32-9bf3-a8813f08762b-config\") pod \"kube-apiserver-operator-766d6c64bb-5dqtb\" (UID: \"27ae76b5-fd48-4e32-9bf3-a8813f08762b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667373 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-trusted-ca\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667398 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dcd0102b-419f-4d81-8187-af734155ca55-auth-proxy-config\") pod \"machine-config-operator-74547568cd-59gxf\" (UID: \"dcd0102b-419f-4d81-8187-af734155ca55\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667427 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx9db\" (UniqueName: \"kubernetes.io/projected/2bf15be5-1380-4b58-934d-7793e22541b4-kube-api-access-qx9db\") pod \"control-plane-machine-set-operator-78cbb6b69f-vw9z9\" (UID: \"2bf15be5-1380-4b58-934d-7793e22541b4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vw9z9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667444 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b50f7b4-dd55-4255-9b68-a942ff4a086e-config-volume\") pod \"dns-default-nzh9r\" (UID: \"7b50f7b4-dd55-4255-9b68-a942ff4a086e\") " pod="openshift-dns/dns-default-nzh9r" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667472 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-console-config\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667490 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2bf15be5-1380-4b58-934d-7793e22541b4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vw9z9\" (UID: \"2bf15be5-1380-4b58-934d-7793e22541b4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vw9z9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667506 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-mountpoint-dir\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667523 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk5pj\" (UniqueName: \"kubernetes.io/projected/b0f5f4f9-02b9-48f1-918e-985b836626ad-kube-api-access-kk5pj\") pod \"ingress-canary-v4fcn\" (UID: \"b0f5f4f9-02b9-48f1-918e-985b836626ad\") " pod="openshift-ingress-canary/ingress-canary-v4fcn" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667561 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/313d7661-5eec-4a28-92a7-cf1f4f85d2e6-config\") pod \"kube-controller-manager-operator-78b949d7b-c4nzd\" (UID: \"313d7661-5eec-4a28-92a7-cf1f4f85d2e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667581 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9a983c06-7704-4c22-8c1b-d0cfcdc67ee8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fgjv4\" (UID: \"9a983c06-7704-4c22-8c1b-d0cfcdc67ee8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fgjv4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667596 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/da5b10e6-e695-4e8a-84e5-329870d5736e-apiservice-cert\") pod \"packageserver-d55dfcdfc-wnnm4\" (UID: \"da5b10e6-e695-4e8a-84e5-329870d5736e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.667616 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk4t2\" (UniqueName: \"kubernetes.io/projected/8cddeb4c-690f-4627-b04a-fedf2fce8594-kube-api-access-nk4t2\") pod \"migrator-59844c95c7-n2cpp\" (UID: \"8cddeb4c-690f-4627-b04a-fedf2fce8594\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n2cpp" Dec 01 17:27:57 crc kubenswrapper[4868]: E1201 17:27:57.668297 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:27:58.168117761 +0000 UTC m=+150.539228172 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.668626 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39ac4687-0a95-43b1-a4b1-334ae97544d2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gkxh9\" (UID: \"39ac4687-0a95-43b1-a4b1-334ae97544d2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.669412 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-oauth-serving-cert\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.670388 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-trusted-ca-bundle\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.672924 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f66e6c7-0b8b-4ef9-b848-ce56e2501193-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljt2x\" (UID: \"1f66e6c7-0b8b-4ef9-b848-ce56e2501193\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.675177 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-service-ca\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.676323 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74782477-7d4a-498c-a3dd-339cebc5ae37-serving-cert\") pod \"service-ca-operator-777779d784-v4hbg\" (UID: \"74782477-7d4a-498c-a3dd-339cebc5ae37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.677097 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-k7ghg\" (UID: \"e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.679601 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-trusted-ca\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.679691 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/313d7661-5eec-4a28-92a7-cf1f4f85d2e6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-c4nzd\" (UID: \"313d7661-5eec-4a28-92a7-cf1f4f85d2e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.680269 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dcd0102b-419f-4d81-8187-af734155ca55-auth-proxy-config\") pod \"machine-config-operator-74547568cd-59gxf\" (UID: \"dcd0102b-419f-4d81-8187-af734155ca55\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.680717 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.681859 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-console-config\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.682045 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-registry-certificates\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.682567 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dcd0102b-419f-4d81-8187-af734155ca55-images\") pod \"machine-config-operator-74547568cd-59gxf\" (UID: \"dcd0102b-419f-4d81-8187-af734155ca55\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.682804 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74782477-7d4a-498c-a3dd-339cebc5ae37-config\") pod \"service-ca-operator-777779d784-v4hbg\" (UID: \"74782477-7d4a-498c-a3dd-339cebc5ae37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.683369 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/313d7661-5eec-4a28-92a7-cf1f4f85d2e6-config\") pod \"kube-controller-manager-operator-78b949d7b-c4nzd\" (UID: \"313d7661-5eec-4a28-92a7-cf1f4f85d2e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.688972 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3bfca371-4e3e-4373-9589-1e1ee7042f85-metrics-tls\") pod \"dns-operator-744455d44c-v8wh9\" (UID: \"3bfca371-4e3e-4373-9589-1e1ee7042f85\") " pod="openshift-dns-operator/dns-operator-744455d44c-v8wh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.689139 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dcd0102b-419f-4d81-8187-af734155ca55-proxy-tls\") pod \"machine-config-operator-74547568cd-59gxf\" (UID: \"dcd0102b-419f-4d81-8187-af734155ca55\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.689402 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f66e6c7-0b8b-4ef9-b848-ce56e2501193-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljt2x\" (UID: \"1f66e6c7-0b8b-4ef9-b848-ce56e2501193\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.692783 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9a983c06-7704-4c22-8c1b-d0cfcdc67ee8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fgjv4\" (UID: \"9a983c06-7704-4c22-8c1b-d0cfcdc67ee8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fgjv4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.705274 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-registry-tls\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.708208 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39ac4687-0a95-43b1-a4b1-334ae97544d2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gkxh9\" (UID: \"39ac4687-0a95-43b1-a4b1-334ae97544d2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.710051 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.712638 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6-proxy-tls\") pod \"machine-config-controller-84d6567774-k7ghg\" (UID: \"e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.716474 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e90813d0-8960-450f-aaa4-047341c50947-console-serving-cert\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.716777 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkkpz\" (UniqueName: \"kubernetes.io/projected/7ca45004-4f76-480b-b8b8-68178408dde2-kube-api-access-bkkpz\") pod \"cluster-samples-operator-665b6dd947-42p77\" (UID: \"7ca45004-4f76-480b-b8b8-68178408dde2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-42p77" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.718791 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/7ca45004-4f76-480b-b8b8-68178408dde2-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-42p77\" (UID: \"7ca45004-4f76-480b-b8b8-68178408dde2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-42p77" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.718982 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e90813d0-8960-450f-aaa4-047341c50947-console-oauth-config\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.788803 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/95e7a05d-5fa9-47ce-8cf3-d261fea05465-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4njnk\" (UID: \"95e7a05d-5fa9-47ce-8cf3-d261fea05465\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.789077 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b0f5f4f9-02b9-48f1-918e-985b836626ad-cert\") pod \"ingress-canary-v4fcn\" (UID: \"b0f5f4f9-02b9-48f1-918e-985b836626ad\") " pod="openshift-ingress-canary/ingress-canary-v4fcn" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.789165 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-plugins-dir\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.789240 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f54tb\" (UniqueName: \"kubernetes.io/projected/95e7a05d-5fa9-47ce-8cf3-d261fea05465-kube-api-access-f54tb\") pod \"package-server-manager-789f6589d5-4njnk\" (UID: \"95e7a05d-5fa9-47ce-8cf3-d261fea05465\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.789339 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.789420 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27ae76b5-fd48-4e32-9bf3-a8813f08762b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5dqtb\" (UID: \"27ae76b5-fd48-4e32-9bf3-a8813f08762b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.789493 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/da5b10e6-e695-4e8a-84e5-329870d5736e-webhook-cert\") pod \"packageserver-d55dfcdfc-wnnm4\" (UID: \"da5b10e6-e695-4e8a-84e5-329870d5736e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.789567 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e3a5f31b-ac9f-4828-82bf-80100560da0c-srv-cert\") pod \"catalog-operator-68c6474976-mt2zl\" (UID: \"e3a5f31b-ac9f-4828-82bf-80100560da0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.789645 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvc2p\" (UniqueName: \"kubernetes.io/projected/b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02-kube-api-access-fvc2p\") pod \"machine-config-server-w5q27\" (UID: \"b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02\") " pod="openshift-machine-config-operator/machine-config-server-w5q27" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.789740 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5e59c27e-823b-4ecf-9ccf-c198528219e9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-m8zhv\" (UID: \"5e59c27e-823b-4ecf-9ccf-c198528219e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.789821 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22cc1879-5e5e-4a06-ba88-f92033382b90-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4hvgn\" (UID: \"22cc1879-5e5e-4a06-ba88-f92033382b90\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.789889 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e59c27e-823b-4ecf-9ccf-c198528219e9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-m8zhv\" (UID: \"5e59c27e-823b-4ecf-9ccf-c198528219e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.789983 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/962e4f14-b280-4ffa-bde0-6cc4b0d86144-config-volume\") pod \"collect-profiles-29410155-bvw5m\" (UID: \"962e4f14-b280-4ffa-bde0-6cc4b0d86144\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.790077 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-registration-dir\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.790164 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7b50f7b4-dd55-4255-9b68-a942ff4a086e-metrics-tls\") pod \"dns-default-nzh9r\" (UID: \"7b50f7b4-dd55-4255-9b68-a942ff4a086e\") " pod="openshift-dns/dns-default-nzh9r" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.790237 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e3a5f31b-ac9f-4828-82bf-80100560da0c-profile-collector-cert\") pod \"catalog-operator-68c6474976-mt2zl\" (UID: \"e3a5f31b-ac9f-4828-82bf-80100560da0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.790319 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/71ff3201-293e-4a9e-ae2a-006c945714ef-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7q4rx\" (UID: \"71ff3201-293e-4a9e-ae2a-006c945714ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.790406 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z625w\" (UniqueName: \"kubernetes.io/projected/71ff3201-293e-4a9e-ae2a-006c945714ef-kube-api-access-z625w\") pod \"olm-operator-6b444d44fb-7q4rx\" (UID: \"71ff3201-293e-4a9e-ae2a-006c945714ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.790512 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/886b276a-b6e7-4ce1-b6cf-dc269d9180bb-stats-auth\") pod \"router-default-5444994796-bddbd\" (UID: \"886b276a-b6e7-4ce1-b6cf-dc269d9180bb\") " pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.790592 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv7k4\" (UniqueName: \"kubernetes.io/projected/e3a5f31b-ac9f-4828-82bf-80100560da0c-kube-api-access-gv7k4\") pod \"catalog-operator-68c6474976-mt2zl\" (UID: \"e3a5f31b-ac9f-4828-82bf-80100560da0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.790664 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm6hg\" (UniqueName: \"kubernetes.io/projected/962e4f14-b280-4ffa-bde0-6cc4b0d86144-kube-api-access-nm6hg\") pod \"collect-profiles-29410155-bvw5m\" (UID: \"962e4f14-b280-4ffa-bde0-6cc4b0d86144\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.790737 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/886b276a-b6e7-4ce1-b6cf-dc269d9180bb-service-ca-bundle\") pod \"router-default-5444994796-bddbd\" (UID: \"886b276a-b6e7-4ce1-b6cf-dc269d9180bb\") " pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.790811 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-socket-dir\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.790881 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/22cc1879-5e5e-4a06-ba88-f92033382b90-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4hvgn\" (UID: \"22cc1879-5e5e-4a06-ba88-f92033382b90\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.790966 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/71ff3201-293e-4a9e-ae2a-006c945714ef-srv-cert\") pod \"olm-operator-6b444d44fb-7q4rx\" (UID: \"71ff3201-293e-4a9e-ae2a-006c945714ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.791164 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/886b276a-b6e7-4ce1-b6cf-dc269d9180bb-metrics-certs\") pod \"router-default-5444994796-bddbd\" (UID: \"886b276a-b6e7-4ce1-b6cf-dc269d9180bb\") " pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.791340 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbf6r\" (UniqueName: \"kubernetes.io/projected/a5bca2c5-996d-49a5-92ac-1d38c96134eb-kube-api-access-zbf6r\") pod \"service-ca-9c57cc56f-q58ld\" (UID: \"a5bca2c5-996d-49a5-92ac-1d38c96134eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-q58ld" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.791437 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27ae76b5-fd48-4e32-9bf3-a8813f08762b-config\") pod \"kube-apiserver-operator-766d6c64bb-5dqtb\" (UID: \"27ae76b5-fd48-4e32-9bf3-a8813f08762b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.791536 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx9db\" (UniqueName: \"kubernetes.io/projected/2bf15be5-1380-4b58-934d-7793e22541b4-kube-api-access-qx9db\") pod \"control-plane-machine-set-operator-78cbb6b69f-vw9z9\" (UID: \"2bf15be5-1380-4b58-934d-7793e22541b4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vw9z9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.791622 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b50f7b4-dd55-4255-9b68-a942ff4a086e-config-volume\") pod \"dns-default-nzh9r\" (UID: \"7b50f7b4-dd55-4255-9b68-a942ff4a086e\") " pod="openshift-dns/dns-default-nzh9r" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.791697 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2bf15be5-1380-4b58-934d-7793e22541b4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vw9z9\" (UID: \"2bf15be5-1380-4b58-934d-7793e22541b4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vw9z9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.791763 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-mountpoint-dir\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.791875 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk5pj\" (UniqueName: \"kubernetes.io/projected/b0f5f4f9-02b9-48f1-918e-985b836626ad-kube-api-access-kk5pj\") pod \"ingress-canary-v4fcn\" (UID: \"b0f5f4f9-02b9-48f1-918e-985b836626ad\") " pod="openshift-ingress-canary/ingress-canary-v4fcn" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.791977 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/da5b10e6-e695-4e8a-84e5-329870d5736e-apiservice-cert\") pod \"packageserver-d55dfcdfc-wnnm4\" (UID: \"da5b10e6-e695-4e8a-84e5-329870d5736e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.792063 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nqj9\" (UniqueName: \"kubernetes.io/projected/22cc1879-5e5e-4a06-ba88-f92033382b90-kube-api-access-9nqj9\") pod \"marketplace-operator-79b997595-4hvgn\" (UID: \"22cc1879-5e5e-4a06-ba88-f92033382b90\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.792134 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/27ae76b5-fd48-4e32-9bf3-a8813f08762b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5dqtb\" (UID: \"27ae76b5-fd48-4e32-9bf3-a8813f08762b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.792223 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a5bca2c5-996d-49a5-92ac-1d38c96134eb-signing-key\") pod \"service-ca-9c57cc56f-q58ld\" (UID: \"a5bca2c5-996d-49a5-92ac-1d38c96134eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-q58ld" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.792504 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzc55\" (UniqueName: \"kubernetes.io/projected/886b276a-b6e7-4ce1-b6cf-dc269d9180bb-kube-api-access-tzc55\") pod \"router-default-5444994796-bddbd\" (UID: \"886b276a-b6e7-4ce1-b6cf-dc269d9180bb\") " pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.792577 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rsjs\" (UniqueName: \"kubernetes.io/projected/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-kube-api-access-2rsjs\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.792664 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m648p\" (UniqueName: \"kubernetes.io/projected/5e59c27e-823b-4ecf-9ccf-c198528219e9-kube-api-access-m648p\") pod \"cluster-image-registry-operator-dc59b4c8b-m8zhv\" (UID: \"5e59c27e-823b-4ecf-9ccf-c198528219e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.792733 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/da5b10e6-e695-4e8a-84e5-329870d5736e-tmpfs\") pod \"packageserver-d55dfcdfc-wnnm4\" (UID: \"da5b10e6-e695-4e8a-84e5-329870d5736e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.792866 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k498r\" (UniqueName: \"kubernetes.io/projected/7b50f7b4-dd55-4255-9b68-a942ff4a086e-kube-api-access-k498r\") pod \"dns-default-nzh9r\" (UID: \"7b50f7b4-dd55-4255-9b68-a942ff4a086e\") " pod="openshift-dns/dns-default-nzh9r" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.792934 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a5bca2c5-996d-49a5-92ac-1d38c96134eb-signing-cabundle\") pod \"service-ca-9c57cc56f-q58ld\" (UID: \"a5bca2c5-996d-49a5-92ac-1d38c96134eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-q58ld" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.793031 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/886b276a-b6e7-4ce1-b6cf-dc269d9180bb-default-certificate\") pod \"router-default-5444994796-bddbd\" (UID: \"886b276a-b6e7-4ce1-b6cf-dc269d9180bb\") " pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.793509 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02-certs\") pod \"machine-config-server-w5q27\" (UID: \"b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02\") " pod="openshift-machine-config-operator/machine-config-server-w5q27" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.793596 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02-node-bootstrap-token\") pod \"machine-config-server-w5q27\" (UID: \"b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02\") " pod="openshift-machine-config-operator/machine-config-server-w5q27" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.793675 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5e59c27e-823b-4ecf-9ccf-c198528219e9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-m8zhv\" (UID: \"5e59c27e-823b-4ecf-9ccf-c198528219e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.793756 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn8pw\" (UniqueName: \"kubernetes.io/projected/da5b10e6-e695-4e8a-84e5-329870d5736e-kube-api-access-kn8pw\") pod \"packageserver-d55dfcdfc-wnnm4\" (UID: \"da5b10e6-e695-4e8a-84e5-329870d5736e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.793842 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/962e4f14-b280-4ffa-bde0-6cc4b0d86144-secret-volume\") pod \"collect-profiles-29410155-bvw5m\" (UID: \"962e4f14-b280-4ffa-bde0-6cc4b0d86144\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.793913 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-csi-data-dir\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.794181 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-csi-data-dir\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.796348 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-plugins-dir\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: E1201 17:27:57.796772 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:27:58.296755575 +0000 UTC m=+150.667865986 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.804763 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27ae76b5-fd48-4e32-9bf3-a8813f08762b-config\") pod \"kube-apiserver-operator-766d6c64bb-5dqtb\" (UID: \"27ae76b5-fd48-4e32-9bf3-a8813f08762b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.804891 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-mountpoint-dir\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.807779 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b50f7b4-dd55-4255-9b68-a942ff4a086e-config-volume\") pod \"dns-default-nzh9r\" (UID: \"7b50f7b4-dd55-4255-9b68-a942ff4a086e\") " pod="openshift-dns/dns-default-nzh9r" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.810085 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/da5b10e6-e695-4e8a-84e5-329870d5736e-tmpfs\") pod \"packageserver-d55dfcdfc-wnnm4\" (UID: \"da5b10e6-e695-4e8a-84e5-329870d5736e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.810880 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a5bca2c5-996d-49a5-92ac-1d38c96134eb-signing-cabundle\") pod \"service-ca-9c57cc56f-q58ld\" (UID: \"a5bca2c5-996d-49a5-92ac-1d38c96134eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-q58ld" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.813704 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/962e4f14-b280-4ffa-bde0-6cc4b0d86144-config-volume\") pod \"collect-profiles-29410155-bvw5m\" (UID: \"962e4f14-b280-4ffa-bde0-6cc4b0d86144\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.815018 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-socket-dir\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.815508 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22cc1879-5e5e-4a06-ba88-f92033382b90-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4hvgn\" (UID: \"22cc1879-5e5e-4a06-ba88-f92033382b90\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.817220 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-registration-dir\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.818690 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcjqp\" (UniqueName: \"kubernetes.io/projected/39ac4687-0a95-43b1-a4b1-334ae97544d2-kube-api-access-dcjqp\") pod \"kube-storage-version-migrator-operator-b67b599dd-gkxh9\" (UID: \"39ac4687-0a95-43b1-a4b1-334ae97544d2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.819470 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b0f5f4f9-02b9-48f1-918e-985b836626ad-cert\") pod \"ingress-canary-v4fcn\" (UID: \"b0f5f4f9-02b9-48f1-918e-985b836626ad\") " pod="openshift-ingress-canary/ingress-canary-v4fcn" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.820131 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/962e4f14-b280-4ffa-bde0-6cc4b0d86144-secret-volume\") pod \"collect-profiles-29410155-bvw5m\" (UID: \"962e4f14-b280-4ffa-bde0-6cc4b0d86144\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.821231 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk4t2\" (UniqueName: \"kubernetes.io/projected/8cddeb4c-690f-4627-b04a-fedf2fce8594-kube-api-access-nk4t2\") pod \"migrator-59844c95c7-n2cpp\" (UID: \"8cddeb4c-690f-4627-b04a-fedf2fce8594\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n2cpp" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.821722 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/886b276a-b6e7-4ce1-b6cf-dc269d9180bb-service-ca-bundle\") pod \"router-default-5444994796-bddbd\" (UID: \"886b276a-b6e7-4ce1-b6cf-dc269d9180bb\") " pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.822595 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e59c27e-823b-4ecf-9ccf-c198528219e9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-m8zhv\" (UID: \"5e59c27e-823b-4ecf-9ccf-c198528219e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.823153 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5e59c27e-823b-4ecf-9ccf-c198528219e9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-m8zhv\" (UID: \"5e59c27e-823b-4ecf-9ccf-c198528219e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.823868 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02-certs\") pod \"machine-config-server-w5q27\" (UID: \"b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02\") " pod="openshift-machine-config-operator/machine-config-server-w5q27" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.826446 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27ae76b5-fd48-4e32-9bf3-a8813f08762b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5dqtb\" (UID: \"27ae76b5-fd48-4e32-9bf3-a8813f08762b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.826472 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e3a5f31b-ac9f-4828-82bf-80100560da0c-srv-cert\") pod \"catalog-operator-68c6474976-mt2zl\" (UID: \"e3a5f31b-ac9f-4828-82bf-80100560da0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.826731 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/886b276a-b6e7-4ce1-b6cf-dc269d9180bb-metrics-certs\") pod \"router-default-5444994796-bddbd\" (UID: \"886b276a-b6e7-4ce1-b6cf-dc269d9180bb\") " pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.827208 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/886b276a-b6e7-4ce1-b6cf-dc269d9180bb-default-certificate\") pod \"router-default-5444994796-bddbd\" (UID: \"886b276a-b6e7-4ce1-b6cf-dc269d9180bb\") " pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.839855 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctll7\" (UniqueName: \"kubernetes.io/projected/74782477-7d4a-498c-a3dd-339cebc5ae37-kube-api-access-ctll7\") pod \"service-ca-operator-777779d784-v4hbg\" (UID: \"74782477-7d4a-498c-a3dd-339cebc5ae37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.841171 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/886b276a-b6e7-4ce1-b6cf-dc269d9180bb-stats-auth\") pod \"router-default-5444994796-bddbd\" (UID: \"886b276a-b6e7-4ce1-b6cf-dc269d9180bb\") " pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.843163 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7b50f7b4-dd55-4255-9b68-a942ff4a086e-metrics-tls\") pod \"dns-default-nzh9r\" (UID: \"7b50f7b4-dd55-4255-9b68-a942ff4a086e\") " pod="openshift-dns/dns-default-nzh9r" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.844214 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/22cc1879-5e5e-4a06-ba88-f92033382b90-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4hvgn\" (UID: \"22cc1879-5e5e-4a06-ba88-f92033382b90\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.845934 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/71ff3201-293e-4a9e-ae2a-006c945714ef-srv-cert\") pod \"olm-operator-6b444d44fb-7q4rx\" (UID: \"71ff3201-293e-4a9e-ae2a-006c945714ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.857862 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02-node-bootstrap-token\") pod \"machine-config-server-w5q27\" (UID: \"b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02\") " pod="openshift-machine-config-operator/machine-config-server-w5q27" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.858599 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.858992 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-828lj\" (UniqueName: \"kubernetes.io/projected/e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6-kube-api-access-828lj\") pod \"machine-config-controller-84d6567774-k7ghg\" (UID: \"e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.859409 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-bound-sa-token\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.860818 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/da5b10e6-e695-4e8a-84e5-329870d5736e-webhook-cert\") pod \"packageserver-d55dfcdfc-wnnm4\" (UID: \"da5b10e6-e695-4e8a-84e5-329870d5736e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.861005 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xdwt\" (UniqueName: \"kubernetes.io/projected/3bfca371-4e3e-4373-9589-1e1ee7042f85-kube-api-access-8xdwt\") pod \"dns-operator-744455d44c-v8wh9\" (UID: \"3bfca371-4e3e-4373-9589-1e1ee7042f85\") " pod="openshift-dns-operator/dns-operator-744455d44c-v8wh9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.865385 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/71ff3201-293e-4a9e-ae2a-006c945714ef-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7q4rx\" (UID: \"71ff3201-293e-4a9e-ae2a-006c945714ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.866645 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a5bca2c5-996d-49a5-92ac-1d38c96134eb-signing-key\") pod \"service-ca-9c57cc56f-q58ld\" (UID: \"a5bca2c5-996d-49a5-92ac-1d38c96134eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-q58ld" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.867302 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/da5b10e6-e695-4e8a-84e5-329870d5736e-apiservice-cert\") pod \"packageserver-d55dfcdfc-wnnm4\" (UID: \"da5b10e6-e695-4e8a-84e5-329870d5736e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.867857 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nwzv\" (UniqueName: \"kubernetes.io/projected/e90813d0-8960-450f-aaa4-047341c50947-kube-api-access-4nwzv\") pod \"console-f9d7485db-c7pwq\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.870420 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bpgd\" (UniqueName: \"kubernetes.io/projected/dcd0102b-419f-4d81-8187-af734155ca55-kube-api-access-9bpgd\") pod \"machine-config-operator-74547568cd-59gxf\" (UID: \"dcd0102b-419f-4d81-8187-af734155ca55\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.872220 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/95e7a05d-5fa9-47ce-8cf3-d261fea05465-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4njnk\" (UID: \"95e7a05d-5fa9-47ce-8cf3-d261fea05465\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.876835 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.878720 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2bf15be5-1380-4b58-934d-7793e22541b4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vw9z9\" (UID: \"2bf15be5-1380-4b58-934d-7793e22541b4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vw9z9" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.884109 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e3a5f31b-ac9f-4828-82bf-80100560da0c-profile-collector-cert\") pod \"catalog-operator-68c6474976-mt2zl\" (UID: \"e3a5f31b-ac9f-4828-82bf-80100560da0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.884378 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.895584 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:27:57 crc kubenswrapper[4868]: E1201 17:27:57.896690 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:27:58.396667615 +0000 UTC m=+150.767778026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.902345 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1f66e6c7-0b8b-4ef9-b848-ce56e2501193-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ljt2x\" (UID: \"1f66e6c7-0b8b-4ef9-b848-ce56e2501193\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.932582 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9klcn\" (UniqueName: \"kubernetes.io/projected/9307075d-5039-4f41-ac9a-8d6c2dbca531-kube-api-access-9klcn\") pod \"downloads-7954f5f757-qnznl\" (UID: \"9307075d-5039-4f41-ac9a-8d6c2dbca531\") " pod="openshift-console/downloads-7954f5f757-qnznl" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.939266 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jflll\" (UniqueName: \"kubernetes.io/projected/9a983c06-7704-4c22-8c1b-d0cfcdc67ee8-kube-api-access-jflll\") pod \"multus-admission-controller-857f4d67dd-fgjv4\" (UID: \"9a983c06-7704-4c22-8c1b-d0cfcdc67ee8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fgjv4" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.949408 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n2cpp" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.951400 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4d64\" (UniqueName: \"kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-kube-api-access-x4d64\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.972327 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-42p77" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.981041 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/313d7661-5eec-4a28-92a7-cf1f4f85d2e6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-c4nzd\" (UID: \"313d7661-5eec-4a28-92a7-cf1f4f85d2e6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd" Dec 01 17:27:57 crc kubenswrapper[4868]: I1201 17:27:57.981779 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn8pw\" (UniqueName: \"kubernetes.io/projected/da5b10e6-e695-4e8a-84e5-329870d5736e-kube-api-access-kn8pw\") pod \"packageserver-d55dfcdfc-wnnm4\" (UID: \"da5b10e6-e695-4e8a-84e5-329870d5736e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.001004 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:58 crc kubenswrapper[4868]: E1201 17:27:58.001644 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:27:58.501410909 +0000 UTC m=+150.872521320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.005790 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f54tb\" (UniqueName: \"kubernetes.io/projected/95e7a05d-5fa9-47ce-8cf3-d261fea05465-kube-api-access-f54tb\") pod \"package-server-manager-789f6589d5-4njnk\" (UID: \"95e7a05d-5fa9-47ce-8cf3-d261fea05465\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.025081 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.040885 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbf6r\" (UniqueName: \"kubernetes.io/projected/a5bca2c5-996d-49a5-92ac-1d38c96134eb-kube-api-access-zbf6r\") pod \"service-ca-9c57cc56f-q58ld\" (UID: \"a5bca2c5-996d-49a5-92ac-1d38c96134eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-q58ld" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.058805 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nqj9\" (UniqueName: \"kubernetes.io/projected/22cc1879-5e5e-4a06-ba88-f92033382b90-kube-api-access-9nqj9\") pod \"marketplace-operator-79b997595-4hvgn\" (UID: \"22cc1879-5e5e-4a06-ba88-f92033382b90\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.062929 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m648p\" (UniqueName: \"kubernetes.io/projected/5e59c27e-823b-4ecf-9ccf-c198528219e9-kube-api-access-m648p\") pod \"cluster-image-registry-operator-dc59b4c8b-m8zhv\" (UID: \"5e59c27e-823b-4ecf-9ccf-c198528219e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.076192 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7q9v5"] Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.076803 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-qnznl" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.082423 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/27ae76b5-fd48-4e32-9bf3-a8813f08762b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5dqtb\" (UID: \"27ae76b5-fd48-4e32-9bf3-a8813f08762b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.101982 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.102913 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk" Dec 01 17:27:58 crc kubenswrapper[4868]: E1201 17:27:58.107485 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:27:58.602667196 +0000 UTC m=+150.973777737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.110023 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.119372 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.135679 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.146489 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx9db\" (UniqueName: \"kubernetes.io/projected/2bf15be5-1380-4b58-934d-7793e22541b4-kube-api-access-qx9db\") pod \"control-plane-machine-set-operator-78cbb6b69f-vw9z9\" (UID: \"2bf15be5-1380-4b58-934d-7793e22541b4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vw9z9" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.147212 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.155911 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-v8wh9" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.168805 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.176796 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk5pj\" (UniqueName: \"kubernetes.io/projected/b0f5f4f9-02b9-48f1-918e-985b836626ad-kube-api-access-kk5pj\") pod \"ingress-canary-v4fcn\" (UID: \"b0f5f4f9-02b9-48f1-918e-985b836626ad\") " pod="openshift-ingress-canary/ingress-canary-v4fcn" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.224092 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:58 crc kubenswrapper[4868]: E1201 17:27:58.224411 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:27:58.72439615 +0000 UTC m=+151.095506561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.226101 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rsjs\" (UniqueName: \"kubernetes.io/projected/d81ae1af-29ea-4b7f-a935-1faa8973fcc9-kube-api-access-2rsjs\") pod \"csi-hostpathplugin-g8kh5\" (UID: \"d81ae1af-29ea-4b7f-a935-1faa8973fcc9\") " pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.230309 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzc55\" (UniqueName: \"kubernetes.io/projected/886b276a-b6e7-4ce1-b6cf-dc269d9180bb-kube-api-access-tzc55\") pod \"router-default-5444994796-bddbd\" (UID: \"886b276a-b6e7-4ce1-b6cf-dc269d9180bb\") " pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.234864 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm6hg\" (UniqueName: \"kubernetes.io/projected/962e4f14-b280-4ffa-bde0-6cc4b0d86144-kube-api-access-nm6hg\") pod \"collect-profiles-29410155-bvw5m\" (UID: \"962e4f14-b280-4ffa-bde0-6cc4b0d86144\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.253478 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fgjv4" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.291578 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5e59c27e-823b-4ecf-9ccf-c198528219e9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-m8zhv\" (UID: \"5e59c27e-823b-4ecf-9ccf-c198528219e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.300617 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k498r\" (UniqueName: \"kubernetes.io/projected/7b50f7b4-dd55-4255-9b68-a942ff4a086e-kube-api-access-k498r\") pod \"dns-default-nzh9r\" (UID: \"7b50f7b4-dd55-4255-9b68-a942ff4a086e\") " pod="openshift-dns/dns-default-nzh9r" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.303654 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.310401 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvc2p\" (UniqueName: \"kubernetes.io/projected/b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02-kube-api-access-fvc2p\") pod \"machine-config-server-w5q27\" (UID: \"b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02\") " pod="openshift-machine-config-operator/machine-config-server-w5q27" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.319565 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.320192 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" event={"ID":"92e9d9e5-3307-46cb-80f8-0a944ba1bab4","Type":"ContainerStarted","Data":"61676d8471a5b57b96eac0c00214ffd1c38830aa082c6a5dbf58be858e337ff2"} Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.338031 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-q58ld" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.339473 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-w5q27" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.340210 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:27:58 crc kubenswrapper[4868]: E1201 17:27:58.342192 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:27:58.842141262 +0000 UTC m=+151.213251673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.352780 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vw9z9" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.352874 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv7k4\" (UniqueName: \"kubernetes.io/projected/e3a5f31b-ac9f-4828-82bf-80100560da0c-kube-api-access-gv7k4\") pod \"catalog-operator-68c6474976-mt2zl\" (UID: \"e3a5f31b-ac9f-4828-82bf-80100560da0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.358856 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z625w\" (UniqueName: \"kubernetes.io/projected/71ff3201-293e-4a9e-ae2a-006c945714ef-kube-api-access-z625w\") pod \"olm-operator-6b444d44fb-7q4rx\" (UID: \"71ff3201-293e-4a9e-ae2a-006c945714ef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.376108 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.391633 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6"] Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.415539 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" event={"ID":"2de90211-84d3-45fc-87eb-efb7d7f16a33","Type":"ContainerStarted","Data":"4feb3f42d3f025037fa114cbca0e420e4b57d8db123f5700f5bc63236989bba2"} Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.416160 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.416393 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.416518 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.421705 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-v4fcn" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.453892 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:58 crc kubenswrapper[4868]: E1201 17:27:58.456262 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:27:58.956235294 +0000 UTC m=+151.327345705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.467458 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.468414 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.489008 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" event={"ID":"221dbb7a-96df-41b4-9c28-240d63013e93","Type":"ContainerStarted","Data":"76cab0c76f093792834b0b658c15d762e000ef1787cec44c1a34fc96b3f7ba5b"} Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.489123 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-262xt"] Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.498288 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.505359 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-nzh9r" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.511921 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.513122 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9"] Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.514517 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" podStartSLOduration=124.514493903 podStartE2EDuration="2m4.514493903s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:58.509362311 +0000 UTC m=+150.880472722" watchObservedRunningTime="2025-12-01 17:27:58.514493903 +0000 UTC m=+150.885604304" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.556400 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:27:58 crc kubenswrapper[4868]: E1201 17:27:58.557312 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:27:59.057252845 +0000 UTC m=+151.428363256 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.657497 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:58 crc kubenswrapper[4868]: E1201 17:27:58.660326 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:27:59.160312282 +0000 UTC m=+151.531422693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.817558 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:27:58 crc kubenswrapper[4868]: E1201 17:27:58.818036 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:27:59.317997059 +0000 UTC m=+151.689107470 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.918926 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:58 crc kubenswrapper[4868]: E1201 17:27:58.919387 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:27:59.419371329 +0000 UTC m=+151.790481740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.966345 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-48kjd" podStartSLOduration=125.966321827 podStartE2EDuration="2m5.966321827s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:58.964878556 +0000 UTC m=+151.335988967" watchObservedRunningTime="2025-12-01 17:27:58.966321827 +0000 UTC m=+151.337432238" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.993869 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" podStartSLOduration=124.993849066 podStartE2EDuration="2m4.993849066s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:58.99035697 +0000 UTC m=+151.361467381" watchObservedRunningTime="2025-12-01 17:27:58.993849066 +0000 UTC m=+151.364959477" Dec 01 17:27:58 crc kubenswrapper[4868]: I1201 17:27:58.995207 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g"] Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.020913 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:27:59 crc kubenswrapper[4868]: E1201 17:27:59.021591 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:27:59.521571372 +0000 UTC m=+151.892681783 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.029608 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qcjvq" podStartSLOduration=126.029581114 podStartE2EDuration="2m6.029581114s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:59.028436642 +0000 UTC m=+151.399547043" watchObservedRunningTime="2025-12-01 17:27:59.029581114 +0000 UTC m=+151.400691525" Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.043554 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf"] Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.147208 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:59 crc kubenswrapper[4868]: E1201 17:27:59.147991 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:27:59.647970654 +0000 UTC m=+152.019081065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.232871 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" podStartSLOduration=126.23285018 podStartE2EDuration="2m6.23285018s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:59.146161415 +0000 UTC m=+151.517271826" watchObservedRunningTime="2025-12-01 17:27:59.23285018 +0000 UTC m=+151.603960591" Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.249115 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:27:59 crc kubenswrapper[4868]: E1201 17:27:59.249299 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:27:59.749273424 +0000 UTC m=+152.120383835 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.249414 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:59 crc kubenswrapper[4868]: E1201 17:27:59.249721 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:27:59.749686325 +0000 UTC m=+152.120796736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.351383 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:27:59 crc kubenswrapper[4868]: E1201 17:27:59.354905 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:27:59.854843219 +0000 UTC m=+152.225953630 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.355593 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:59 crc kubenswrapper[4868]: E1201 17:27:59.356186 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:27:59.856176487 +0000 UTC m=+152.227286898 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.457650 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:27:59 crc kubenswrapper[4868]: E1201 17:27:59.458069 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:27:59.958048901 +0000 UTC m=+152.329159312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.497527 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" event={"ID":"caa5fa90-31b2-49f5-954e-9640fdd8140d","Type":"ContainerStarted","Data":"11a12ae7c890fc609edfb92fb3ed875793b2c69e0262c2bd8c131f23e8e87e09"} Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.497955 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-qw86w" podStartSLOduration=125.497933422 podStartE2EDuration="2m5.497933422s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:59.465727453 +0000 UTC m=+151.836837864" watchObservedRunningTime="2025-12-01 17:27:59.497933422 +0000 UTC m=+151.869043833" Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.509871 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bddbd" event={"ID":"886b276a-b6e7-4ce1-b6cf-dc269d9180bb","Type":"ContainerStarted","Data":"fe1c868d52b6fc6c11d23c2ac5fc2ae04b1cdf1212e5bbe41f46166a0adab3f2"} Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.514368 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" event={"ID":"e8306af7-7828-4bf6-8099-3a986fd2572b","Type":"ContainerStarted","Data":"9e53401aa1a6da47e401a3845a4a8d13f13fc71629fd9a6e4215b5137787d333"} Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.515458 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g" event={"ID":"482c868b-ee7d-48d9-b9fb-8fb0527c95bf","Type":"ContainerStarted","Data":"f43ff8b5bd4348b6ca216cea79568a09c28b432646679d5654a042a3df4ac155"} Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.519435 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" event={"ID":"221dbb7a-96df-41b4-9c28-240d63013e93","Type":"ContainerStarted","Data":"2d754ca6cf64a892bf524ad0eeff588c45368823ae2dea34a6c1eaa42a042c09"} Dec 01 17:27:59 crc kubenswrapper[4868]: W1201 17:27:59.520446 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcd0102b_419f_4d81_8187_af734155ca55.slice/crio-afe255c01ca5faf984934232b1ca455182a18fc0d4ad75d516b12118281f6425 WatchSource:0}: Error finding container afe255c01ca5faf984934232b1ca455182a18fc0d4ad75d516b12118281f6425: Status 404 returned error can't find the container with id afe255c01ca5faf984934232b1ca455182a18fc0d4ad75d516b12118281f6425 Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.521135 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-w5q27" event={"ID":"b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02","Type":"ContainerStarted","Data":"a9fee52ba25d96d69e671eba7907fafcdebcd65d26568f433df8a6e44a5ab459"} Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.536603 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-262xt" event={"ID":"619b567a-92a9-4801-991e-792545b45bd0","Type":"ContainerStarted","Data":"203f5fb8ceb64254eba919d4792e7456de845c5ef977ce11a0a5f65cce50a407"} Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.551441 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9" event={"ID":"39ac4687-0a95-43b1-a4b1-334ae97544d2","Type":"ContainerStarted","Data":"bc88a704769565a13710d0ab1eeea13c40af391d81e17459629df742aad84d58"} Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.565428 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:59 crc kubenswrapper[4868]: E1201 17:27:59.566161 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:00.066145558 +0000 UTC m=+152.437255969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.666740 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:27:59 crc kubenswrapper[4868]: E1201 17:27:59.680219 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:00.180184327 +0000 UTC m=+152.551294918 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.728723 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" podStartSLOduration=126.728694368 podStartE2EDuration="2m6.728694368s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:27:59.669252565 +0000 UTC m=+152.040362976" watchObservedRunningTime="2025-12-01 17:27:59.728694368 +0000 UTC m=+152.099804789" Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.773807 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:59 crc kubenswrapper[4868]: E1201 17:27:59.774177 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:00.274163244 +0000 UTC m=+152.645273655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.875691 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:27:59 crc kubenswrapper[4868]: E1201 17:27:59.876095 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:00.376076279 +0000 UTC m=+152.747186690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:27:59 crc kubenswrapper[4868]: I1201 17:27:59.992238 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:27:59 crc kubenswrapper[4868]: E1201 17:27:59.993213 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:00.493195585 +0000 UTC m=+152.864305996 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.093435 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:00 crc kubenswrapper[4868]: E1201 17:28:00.093668 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:00.593620169 +0000 UTC m=+152.964730580 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.094105 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:00 crc kubenswrapper[4868]: E1201 17:28:00.094498 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:00.594483293 +0000 UTC m=+152.965593704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.198349 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:00 crc kubenswrapper[4868]: E1201 17:28:00.198928 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:00.698902809 +0000 UTC m=+153.070013220 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.216224 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" podStartSLOduration=127.216199796 podStartE2EDuration="2m7.216199796s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:00.197605652 +0000 UTC m=+152.568716063" watchObservedRunningTime="2025-12-01 17:28:00.216199796 +0000 UTC m=+152.587310207" Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.300774 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:00 crc kubenswrapper[4868]: E1201 17:28:00.301563 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:00.801547644 +0000 UTC m=+153.172658055 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.403779 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:00 crc kubenswrapper[4868]: E1201 17:28:00.404739 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:00.904714764 +0000 UTC m=+153.275825175 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.409310 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.409811 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.425406 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-pvqvn" podStartSLOduration=127.425382225 podStartE2EDuration="2m7.425382225s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:00.420827559 +0000 UTC m=+152.791937970" watchObservedRunningTime="2025-12-01 17:28:00.425382225 +0000 UTC m=+152.796492636" Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.450249 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.450739 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.466705 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.502817 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-n2cpp"] Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.505815 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:00 crc kubenswrapper[4868]: E1201 17:28:00.506768 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:01.006753423 +0000 UTC m=+153.377863834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.565013 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-c7pwq"] Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.569386 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-w5q27" event={"ID":"b9fbef6b-eee2-40a4-a2b6-8eba86f5fd02","Type":"ContainerStarted","Data":"65890a021fce7cad6595170d76645555171f692ff04f25995fc6a2e68fc199d1"} Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.574463 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-262xt" event={"ID":"619b567a-92a9-4801-991e-792545b45bd0","Type":"ContainerStarted","Data":"eb516413dbb8103dfed9b096c454c28ac6eb0674f0e873689373021409a213a7"} Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.574966 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.576008 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9" event={"ID":"39ac4687-0a95-43b1-a4b1-334ae97544d2","Type":"ContainerStarted","Data":"353379d01584816809cf37b9f5db446d6a7ec73d934aab5ac25b22ddd8efa9f5"} Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.577258 4868 patch_prober.go:28] interesting pod/console-operator-58897d9998-262xt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.577318 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-262xt" podUID="619b567a-92a9-4801-991e-792545b45bd0" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.580976 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" event={"ID":"caa5fa90-31b2-49f5-954e-9640fdd8140d","Type":"ContainerStarted","Data":"f966497f544a8eba95e38e0571fbd6793be55496d65f355b205bf06da12b6970"} Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.582463 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bddbd" event={"ID":"886b276a-b6e7-4ce1-b6cf-dc269d9180bb","Type":"ContainerStarted","Data":"a40e0a6e76c501515ca645580d5e9bd1d86f7b31fa40f8b6030cb6a00fc5204a"} Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.583557 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" event={"ID":"e8306af7-7828-4bf6-8099-3a986fd2572b","Type":"ContainerStarted","Data":"9c49f766f49d9b2cc89ff42e5e66992d74af044fa61c3eb580196e369e0ad129"} Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.585982 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g" event={"ID":"482c868b-ee7d-48d9-b9fb-8fb0527c95bf","Type":"ContainerStarted","Data":"5824d4ccf8522fc4e91fd41c5eded6b005b8c0ef69dfb241496d3a5715e7a9be"} Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.589037 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" event={"ID":"dcd0102b-419f-4d81-8187-af734155ca55","Type":"ContainerStarted","Data":"32dc84d59437d797afd7896fca73d507772b9655b6b41a12af3737193b8d55d4"} Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.589119 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" event={"ID":"dcd0102b-419f-4d81-8187-af734155ca55","Type":"ContainerStarted","Data":"afe255c01ca5faf984934232b1ca455182a18fc0d4ad75d516b12118281f6425"} Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.611687 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:00 crc kubenswrapper[4868]: E1201 17:28:00.617894 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:01.117860083 +0000 UTC m=+153.488970494 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.639730 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-62tk9" Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.660974 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" podStartSLOduration=127.660922932 podStartE2EDuration="2m7.660922932s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:00.605508792 +0000 UTC m=+152.976619223" watchObservedRunningTime="2025-12-01 17:28:00.660922932 +0000 UTC m=+153.032033343" Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.719508 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:00 crc kubenswrapper[4868]: E1201 17:28:00.721419 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:01.221401903 +0000 UTC m=+153.592512324 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.798988 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-bddbd" podStartSLOduration=126.798961096 podStartE2EDuration="2m6.798961096s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:00.719880332 +0000 UTC m=+153.090990743" watchObservedRunningTime="2025-12-01 17:28:00.798961096 +0000 UTC m=+153.170071507" Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.821471 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:00 crc kubenswrapper[4868]: E1201 17:28:00.822191 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:01.322167297 +0000 UTC m=+153.693277708 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.920934 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gkxh9" podStartSLOduration=126.920894804 podStartE2EDuration="2m6.920894804s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:00.917683906 +0000 UTC m=+153.288794327" watchObservedRunningTime="2025-12-01 17:28:00.920894804 +0000 UTC m=+153.292005215" Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.922641 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-262xt" podStartSLOduration=127.922630952 podStartE2EDuration="2m7.922630952s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:00.884304793 +0000 UTC m=+153.255415204" watchObservedRunningTime="2025-12-01 17:28:00.922630952 +0000 UTC m=+153.293741363" Dec 01 17:28:00 crc kubenswrapper[4868]: I1201 17:28:00.926817 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:00 crc kubenswrapper[4868]: E1201 17:28:00.927243 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:01.42722993 +0000 UTC m=+153.798340341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.003569 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-7q9v5" podStartSLOduration=128.003540598 podStartE2EDuration="2m8.003540598s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:00.983854934 +0000 UTC m=+153.354965345" watchObservedRunningTime="2025-12-01 17:28:01.003540598 +0000 UTC m=+153.374651009" Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.029579 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:01 crc kubenswrapper[4868]: E1201 17:28:01.030345 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:01.530327077 +0000 UTC m=+153.901437488 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.144686 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:01 crc kubenswrapper[4868]: E1201 17:28:01.145476 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:01.645461598 +0000 UTC m=+154.016572009 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.249749 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:01 crc kubenswrapper[4868]: E1201 17:28:01.250801 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:01.750774148 +0000 UTC m=+154.121884569 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.341354 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-w5q27" podStartSLOduration=6.341331319 podStartE2EDuration="6.341331319s" podCreationTimestamp="2025-12-01 17:27:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:01.339316224 +0000 UTC m=+153.710426635" watchObservedRunningTime="2025-12-01 17:28:01.341331319 +0000 UTC m=+153.712441730" Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.360099 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:01 crc kubenswrapper[4868]: E1201 17:28:01.360725 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:01.860706775 +0000 UTC m=+154.231817186 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.400008 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pwq5g" podStartSLOduration=128.39998568 podStartE2EDuration="2m8.39998568s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:01.393359617 +0000 UTC m=+153.764470018" watchObservedRunningTime="2025-12-01 17:28:01.39998568 +0000 UTC m=+153.771096081" Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.463448 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:01 crc kubenswrapper[4868]: E1201 17:28:01.463722 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:01.96368707 +0000 UTC m=+154.334797481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.464518 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:01 crc kubenswrapper[4868]: E1201 17:28:01.465051 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:01.965035378 +0000 UTC m=+154.336145789 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.473527 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.473728 4868 patch_prober.go:28] interesting pod/router-default-5444994796-bddbd container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.473767 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bddbd" podUID="886b276a-b6e7-4ce1-b6cf-dc269d9180bb" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.568810 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:01 crc kubenswrapper[4868]: E1201 17:28:01.568955 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:02.068909937 +0000 UTC m=+154.440020348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.569363 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:01 crc kubenswrapper[4868]: E1201 17:28:01.569838 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:02.069827082 +0000 UTC m=+154.440937493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.588847 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vr48t" Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.613381 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" event={"ID":"dcd0102b-419f-4d81-8187-af734155ca55","Type":"ContainerStarted","Data":"af9e481c6810be27ecd6bef9f5bef7b58e5bd6ee2556aa64ae51c9c889e7c1f2"} Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.625411 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n2cpp" event={"ID":"8cddeb4c-690f-4627-b04a-fedf2fce8594","Type":"ContainerStarted","Data":"ea43c6e1264342857dd57452cbe9b4b7e84ca0d35eb32768970076a4db9958ea"} Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.625467 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n2cpp" event={"ID":"8cddeb4c-690f-4627-b04a-fedf2fce8594","Type":"ContainerStarted","Data":"b8f2b7c535500528c341c39de961e29188b1823e3e01979fa6f7d7b2f771af59"} Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.629296 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk"] Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.655012 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" event={"ID":"e8306af7-7828-4bf6-8099-3a986fd2572b","Type":"ContainerStarted","Data":"1f40a56cc21d95bce01fd4feb5e54001d9cee96c0e520bf5584c44579fdf219d"} Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.670307 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.672032 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4hvgn"] Dec 01 17:28:01 crc kubenswrapper[4868]: E1201 17:28:01.674142 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:02.174105173 +0000 UTC m=+154.545215584 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.757982 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-c7pwq" event={"ID":"e90813d0-8960-450f-aaa4-047341c50947","Type":"ContainerStarted","Data":"76f9542b3855fd35afaa09389ce52b6759951098da66cf38e2857b1c1414008f"} Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.758035 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-c7pwq" event={"ID":"e90813d0-8960-450f-aaa4-047341c50947","Type":"ContainerStarted","Data":"f57bb897b63f0c317a548208464081e0b32a139de0e8639de7432b9b76808f2c"} Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.766005 4868 patch_prober.go:28] interesting pod/console-operator-58897d9998-262xt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.766111 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-262xt" podUID="619b567a-92a9-4801-991e-792545b45bd0" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.795304 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:01 crc kubenswrapper[4868]: E1201 17:28:01.795732 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:02.295717103 +0000 UTC m=+154.666827514 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.884522 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x"] Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.885204 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4hl6" podStartSLOduration=128.885189384 podStartE2EDuration="2m8.885189384s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:01.816661471 +0000 UTC m=+154.187771882" watchObservedRunningTime="2025-12-01 17:28:01.885189384 +0000 UTC m=+154.256299795" Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.921753 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-59gxf" podStartSLOduration=127.921720404 podStartE2EDuration="2m7.921720404s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:01.898144723 +0000 UTC m=+154.269255144" watchObservedRunningTime="2025-12-01 17:28:01.921720404 +0000 UTC m=+154.292830815" Dec 01 17:28:01 crc kubenswrapper[4868]: E1201 17:28:01.984281 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:02.484242471 +0000 UTC m=+154.855352882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.961004 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.991626 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:01 crc kubenswrapper[4868]: I1201 17:28:01.993321 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-42p77"] Dec 01 17:28:01 crc kubenswrapper[4868]: E1201 17:28:01.993488 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:02.493471556 +0000 UTC m=+154.864581957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.014147 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.023678 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.039740 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-c7pwq" podStartSLOduration=129.039713434 podStartE2EDuration="2m9.039713434s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:01.984770376 +0000 UTC m=+154.355880787" watchObservedRunningTime="2025-12-01 17:28:02.039713434 +0000 UTC m=+154.410823845" Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.094008 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-qnznl"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.094073 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.094858 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:02 crc kubenswrapper[4868]: E1201 17:28:02.095319 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:02.595297619 +0000 UTC m=+154.966408030 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.115331 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fgjv4"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.135409 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vw9z9"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.153800 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.160352 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.164619 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv"] Dec 01 17:28:02 crc kubenswrapper[4868]: W1201 17:28:02.165076 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bf15be5_1380_4b58_934d_7793e22541b4.slice/crio-3eb40f5061b5bf98d37ca7541503ad47ec7116e1e31c79fee3c03f55ce6cfeba WatchSource:0}: Error finding container 3eb40f5061b5bf98d37ca7541503ad47ec7116e1e31c79fee3c03f55ce6cfeba: Status 404 returned error can't find the container with id 3eb40f5061b5bf98d37ca7541503ad47ec7116e1e31c79fee3c03f55ce6cfeba Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.170104 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.179857 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-v8wh9"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.179899 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.197628 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:02 crc kubenswrapper[4868]: E1201 17:28:02.198362 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:02.698342866 +0000 UTC m=+155.069453277 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.202334 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.300728 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:02 crc kubenswrapper[4868]: E1201 17:28:02.301066 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:02.801048553 +0000 UTC m=+155.172158964 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.320629 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-nzh9r"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.328040 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-v4fcn"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.335395 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q58ld"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.370437 4868 patch_prober.go:28] interesting pod/apiserver-76f77b778f-hwgjb container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 01 17:28:02 crc kubenswrapper[4868]: [+]log ok Dec 01 17:28:02 crc kubenswrapper[4868]: [+]etcd ok Dec 01 17:28:02 crc kubenswrapper[4868]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 01 17:28:02 crc kubenswrapper[4868]: [+]poststarthook/generic-apiserver-start-informers ok Dec 01 17:28:02 crc kubenswrapper[4868]: [+]poststarthook/max-in-flight-filter ok Dec 01 17:28:02 crc kubenswrapper[4868]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 01 17:28:02 crc kubenswrapper[4868]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 01 17:28:02 crc kubenswrapper[4868]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 01 17:28:02 crc kubenswrapper[4868]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 01 17:28:02 crc kubenswrapper[4868]: [+]poststarthook/project.openshift.io-projectcache ok Dec 01 17:28:02 crc kubenswrapper[4868]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 01 17:28:02 crc kubenswrapper[4868]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Dec 01 17:28:02 crc kubenswrapper[4868]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 01 17:28:02 crc kubenswrapper[4868]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 01 17:28:02 crc kubenswrapper[4868]: livez check failed Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.370531 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" podUID="221dbb7a-96df-41b4-9c28-240d63013e93" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.402958 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:02 crc kubenswrapper[4868]: E1201 17:28:02.403332 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:02.903317828 +0000 UTC m=+155.274428229 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.467408 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-g8kh5"] Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.478225 4868 patch_prober.go:28] interesting pod/router-default-5444994796-bddbd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 17:28:02 crc kubenswrapper[4868]: [-]has-synced failed: reason withheld Dec 01 17:28:02 crc kubenswrapper[4868]: [+]process-running ok Dec 01 17:28:02 crc kubenswrapper[4868]: healthz check failed Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.478307 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bddbd" podUID="886b276a-b6e7-4ce1-b6cf-dc269d9180bb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.504644 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:02 crc kubenswrapper[4868]: E1201 17:28:02.505171 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:03.005149082 +0000 UTC m=+155.376259493 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.606001 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:02 crc kubenswrapper[4868]: E1201 17:28:02.609085 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:03.109064623 +0000 UTC m=+155.480175034 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.710606 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:02 crc kubenswrapper[4868]: E1201 17:28:02.711050 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:03.21103133 +0000 UTC m=+155.582141741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.763520 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qnznl" event={"ID":"9307075d-5039-4f41-ac9a-8d6c2dbca531","Type":"ContainerStarted","Data":"8be494799fa1ca47fef75ee37d25f33c1f9061e74adb238d211d708d8e866faf"} Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.763585 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qnznl" event={"ID":"9307075d-5039-4f41-ac9a-8d6c2dbca531","Type":"ContainerStarted","Data":"1066e0e46e4f175040cd50171644f90ae5ef0925188759538b35d9c7ae274cd1"} Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.764520 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-qnznl" Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.768126 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.768198 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.812144 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:02 crc kubenswrapper[4868]: E1201 17:28:02.812660 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:03.312643707 +0000 UTC m=+155.683754118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.857745 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-qnznl" podStartSLOduration=129.857718293 podStartE2EDuration="2m9.857718293s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:02.857313042 +0000 UTC m=+155.228423453" watchObservedRunningTime="2025-12-01 17:28:02.857718293 +0000 UTC m=+155.228828704" Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.922123 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.922854 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg" event={"ID":"74782477-7d4a-498c-a3dd-339cebc5ae37","Type":"ContainerStarted","Data":"7703ffdcb4b92c6b2396002fffeb3593caceb9306ee79f109b8f421002f8d395"} Dec 01 17:28:02 crc kubenswrapper[4868]: E1201 17:28:02.924542 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:03.424497857 +0000 UTC m=+155.795608268 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:02 crc kubenswrapper[4868]: I1201 17:28:02.958510 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" event={"ID":"d81ae1af-29ea-4b7f-a935-1faa8973fcc9","Type":"ContainerStarted","Data":"893a671fc18fe69e266be1e7c8149deef86bef35fc7dab464c23f47b858533b0"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.001705 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" event={"ID":"962e4f14-b280-4ffa-bde0-6cc4b0d86144","Type":"ContainerStarted","Data":"1a1ddd52c91cd0d66eb9ce648f30ab72b5598fdd4aa147bf1505ead58516cba8"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.003080 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fgjv4" event={"ID":"9a983c06-7704-4c22-8c1b-d0cfcdc67ee8","Type":"ContainerStarted","Data":"bd80b53a6d4a9401ca4a939592aa6981422e846035506805e79e54ee2b250075"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.031895 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:03 crc kubenswrapper[4868]: E1201 17:28:03.032529 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:03.532516671 +0000 UTC m=+155.903627082 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.076658 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd" event={"ID":"313d7661-5eec-4a28-92a7-cf1f4f85d2e6","Type":"ContainerStarted","Data":"a47ead5cc42e6ee7aada05e2960ab3a81a41be6d9a1ffd9b17108012b3afca21"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.107207 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" event={"ID":"e3a5f31b-ac9f-4828-82bf-80100560da0c","Type":"ContainerStarted","Data":"9ab9bafd2014fa83dad6401f91e91d3b552bded7c1a68f7d750ce9f92d1509f3"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.107299 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" event={"ID":"e3a5f31b-ac9f-4828-82bf-80100560da0c","Type":"ContainerStarted","Data":"069bc58ac41c12cae73f7324bf6624c79030ee50968609581d71159946cbfe0c"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.107428 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.109910 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-q58ld" event={"ID":"a5bca2c5-996d-49a5-92ac-1d38c96134eb","Type":"ContainerStarted","Data":"53e9d2968bf9c73731362e3ecfbb56f16620a2e608dd69f7a7374c7990c0bf93"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.113533 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb" event={"ID":"27ae76b5-fd48-4e32-9bf3-a8813f08762b","Type":"ContainerStarted","Data":"b2e38a8e52e9cba18d9009049997ed2e0cfbdff6a6c5508cd88d1f3b254c112d"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.120081 4868 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-mt2zl container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.120144 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" podUID="e3a5f31b-ac9f-4828-82bf-80100560da0c" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.150076 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:03 crc kubenswrapper[4868]: E1201 17:28:03.150760 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:03.65048128 +0000 UTC m=+156.021591691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.150806 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:03 crc kubenswrapper[4868]: E1201 17:28:03.151525 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:03.651513099 +0000 UTC m=+156.022623510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.175710 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vw9z9" event={"ID":"2bf15be5-1380-4b58-934d-7793e22541b4","Type":"ContainerStarted","Data":"3eb40f5061b5bf98d37ca7541503ad47ec7116e1e31c79fee3c03f55ce6cfeba"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.254184 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:03 crc kubenswrapper[4868]: E1201 17:28:03.254994 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:03.754958856 +0000 UTC m=+156.126069267 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.329664 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x" event={"ID":"1f66e6c7-0b8b-4ef9-b848-ce56e2501193","Type":"ContainerStarted","Data":"dfc8fdad49a813b2ce7a4582101d3c407368a84c889da011c274d43d1b57de67"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.329722 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x" event={"ID":"1f66e6c7-0b8b-4ef9-b848-ce56e2501193","Type":"ContainerStarted","Data":"5a0d6600b2c3b1aa5dd411887a10b9d446fc622f8ab76a4530ec1c88978c37da"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.356892 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:03 crc kubenswrapper[4868]: E1201 17:28:03.357353 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:03.857338735 +0000 UTC m=+156.228449136 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.357681 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n2cpp" event={"ID":"8cddeb4c-690f-4627-b04a-fedf2fce8594","Type":"ContainerStarted","Data":"39020c93865f9999b375333ddf88cd406d6d4559aacf2daed2321bb0ddddc7e2"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.369683 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ljt2x" podStartSLOduration=129.369664285 podStartE2EDuration="2m9.369664285s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:03.368257307 +0000 UTC m=+155.739367718" watchObservedRunningTime="2025-12-01 17:28:03.369664285 +0000 UTC m=+155.740774696" Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.369839 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" podStartSLOduration=129.36983495 podStartE2EDuration="2m9.36983495s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:03.28801043 +0000 UTC m=+155.659120841" watchObservedRunningTime="2025-12-01 17:28:03.36983495 +0000 UTC m=+155.740945351" Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.393169 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk" event={"ID":"95e7a05d-5fa9-47ce-8cf3-d261fea05465","Type":"ContainerStarted","Data":"bb5c779044b8207edf3df85164c48acf949593015c896a410bda1155ac0d4061"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.393229 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk" event={"ID":"95e7a05d-5fa9-47ce-8cf3-d261fea05465","Type":"ContainerStarted","Data":"2c77fc22c33f6491a5c4b8001d9d54a9b864a7de71a5494f97bb4a16fb66636e"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.394441 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" event={"ID":"22cc1879-5e5e-4a06-ba88-f92033382b90","Type":"ContainerStarted","Data":"8e73eb20d2657fbf2df84c444f2df5c0d7a00fddee5dea5a79f5697a15825991"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.394472 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" event={"ID":"22cc1879-5e5e-4a06-ba88-f92033382b90","Type":"ContainerStarted","Data":"bbc91beeabbd54e17c194ad29b3c73dde0a7a4b3f01415f83582bb754891b950"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.395573 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.433640 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-42p77" event={"ID":"7ca45004-4f76-480b-b8b8-68178408dde2","Type":"ContainerStarted","Data":"a76f7dfbb6713b37e193302b703d72b1a8ffee38a5b649f12615e1acadf61ce9"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.436818 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n2cpp" podStartSLOduration=129.43680516 podStartE2EDuration="2m9.43680516s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:03.429201351 +0000 UTC m=+155.800311782" watchObservedRunningTime="2025-12-01 17:28:03.43680516 +0000 UTC m=+155.807915571" Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.438880 4868 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4hvgn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.438936 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" podUID="22cc1879-5e5e-4a06-ba88-f92033382b90" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.459098 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:03 crc kubenswrapper[4868]: E1201 17:28:03.460802 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:03.960784453 +0000 UTC m=+156.331894864 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.471707 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" podStartSLOduration=129.471677474 podStartE2EDuration="2m9.471677474s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:03.465090091 +0000 UTC m=+155.836200502" watchObservedRunningTime="2025-12-01 17:28:03.471677474 +0000 UTC m=+155.842787975" Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.479484 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-v8wh9" event={"ID":"3bfca371-4e3e-4373-9589-1e1ee7042f85","Type":"ContainerStarted","Data":"34cb5ce0fa00477aec14952e3648c229079415d47d346f341b0695fa088215dc"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.485094 4868 patch_prober.go:28] interesting pod/router-default-5444994796-bddbd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 17:28:03 crc kubenswrapper[4868]: [-]has-synced failed: reason withheld Dec 01 17:28:03 crc kubenswrapper[4868]: [+]process-running ok Dec 01 17:28:03 crc kubenswrapper[4868]: healthz check failed Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.485184 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bddbd" podUID="886b276a-b6e7-4ce1-b6cf-dc269d9180bb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.507623 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-nzh9r" event={"ID":"7b50f7b4-dd55-4255-9b68-a942ff4a086e","Type":"ContainerStarted","Data":"b60137d32b6ce7ce8b83fa3ed95b3d38320ee6f0fca10c2e954057ca08872c97"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.512044 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-v4fcn" event={"ID":"b0f5f4f9-02b9-48f1-918e-985b836626ad","Type":"ContainerStarted","Data":"fec3e62bcfce5f92f05f004effd048db954d7afb89dbc4509665faa3c31f2d7b"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.515116 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" event={"ID":"da5b10e6-e695-4e8a-84e5-329870d5736e","Type":"ContainerStarted","Data":"09ba19e758e1bdf1b497b3901b242f35affd2922ef69a1b9d033db68f5d9d381"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.516002 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.565015 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:03 crc kubenswrapper[4868]: E1201 17:28:03.566609 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:04.066593786 +0000 UTC m=+156.437704197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.577310 4868 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wnnm4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.577374 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" podUID="da5b10e6-e695-4e8a-84e5-329870d5736e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.638588 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg" event={"ID":"e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6","Type":"ContainerStarted","Data":"09a2ca40191da3281e405e7aa421ac360b0f946d60b0326bd5205c99e7cbd66c"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.660818 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" event={"ID":"71ff3201-293e-4a9e-ae2a-006c945714ef","Type":"ContainerStarted","Data":"f91e71d180d2158fa1d706bfe2393d8eb5c660374cb248136f598ccfc7d3725e"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.669657 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:03 crc kubenswrapper[4868]: E1201 17:28:03.670718 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:04.170700862 +0000 UTC m=+156.541811273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.676800 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" event={"ID":"5e59c27e-823b-4ecf-9ccf-c198528219e9","Type":"ContainerStarted","Data":"088347a0b44c01673413542242f30bc8dcd9e4a2fcb0f85da60a75543e2a8718"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.676841 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" event={"ID":"5e59c27e-823b-4ecf-9ccf-c198528219e9","Type":"ContainerStarted","Data":"fa84ca6a819dd411bfa35c0512c48aea63612dcb79389a960ce78f077322c280"} Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.721360 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" podStartSLOduration=129.721336951 podStartE2EDuration="2m9.721336951s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:03.658495484 +0000 UTC m=+156.029605895" watchObservedRunningTime="2025-12-01 17:28:03.721336951 +0000 UTC m=+156.092447362" Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.773853 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:03 crc kubenswrapper[4868]: E1201 17:28:03.777302 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:04.277282177 +0000 UTC m=+156.648392588 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:03 crc kubenswrapper[4868]: I1201 17:28:03.878217 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:03 crc kubenswrapper[4868]: E1201 17:28:03.878615 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:04.378596666 +0000 UTC m=+156.749707077 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:03.985541 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:04 crc kubenswrapper[4868]: E1201 17:28:03.985899 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:04.485886059 +0000 UTC m=+156.856996470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.098524 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:04 crc kubenswrapper[4868]: E1201 17:28:04.099129 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:04.599109638 +0000 UTC m=+156.970220049 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.201088 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:04 crc kubenswrapper[4868]: E1201 17:28:04.202012 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:04.70199295 +0000 UTC m=+157.073103361 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.323490 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:04 crc kubenswrapper[4868]: E1201 17:28:04.323881 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:04.823861256 +0000 UTC m=+157.194971667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.428016 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:04 crc kubenswrapper[4868]: E1201 17:28:04.428556 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:04.928532878 +0000 UTC m=+157.299643289 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.485284 4868 patch_prober.go:28] interesting pod/router-default-5444994796-bddbd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 17:28:04 crc kubenswrapper[4868]: [-]has-synced failed: reason withheld Dec 01 17:28:04 crc kubenswrapper[4868]: [+]process-running ok Dec 01 17:28:04 crc kubenswrapper[4868]: healthz check failed Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.485822 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bddbd" podUID="886b276a-b6e7-4ce1-b6cf-dc269d9180bb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.533135 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:04 crc kubenswrapper[4868]: E1201 17:28:04.533396 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:05.033378685 +0000 UTC m=+157.404489096 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.635705 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:04 crc kubenswrapper[4868]: E1201 17:28:04.636164 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:05.136147833 +0000 UTC m=+157.507258244 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.695189 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg" event={"ID":"74782477-7d4a-498c-a3dd-339cebc5ae37","Type":"ContainerStarted","Data":"ad00b7112f242625911372d33398b2fe76a99837b1a7c89513f045301bd7fbfa"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.697851 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" event={"ID":"962e4f14-b280-4ffa-bde0-6cc4b0d86144","Type":"ContainerStarted","Data":"83cbd781a24db21510be592699a361e90e624e963509bd92bbef568a0b050db1"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.700848 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-nzh9r" event={"ID":"7b50f7b4-dd55-4255-9b68-a942ff4a086e","Type":"ContainerStarted","Data":"a87d7ee99e70cd08a38e845008e7dbf56f6080b03ed45e93a8850928f0bb2657"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.702107 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd" event={"ID":"313d7661-5eec-4a28-92a7-cf1f4f85d2e6","Type":"ContainerStarted","Data":"4603921ce0991701488a16710cb0cd1bec0dd6e3fc79c5a44e2d7fe57aa49a03"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.705250 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg" event={"ID":"e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6","Type":"ContainerStarted","Data":"12e25f10bff21f124185d7cdf52264a3cfe151fb8a4b5d6d988062bb808539f2"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.705289 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg" event={"ID":"e4e72efa-9ad8-42f7-a9cf-9967ca8b47f6","Type":"ContainerStarted","Data":"3d9fa6491e03c6f2fd2592cafac6a5b8bd09a033d8c4190b7b4beaa99b19d6a1"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.709040 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb" event={"ID":"27ae76b5-fd48-4e32-9bf3-a8813f08762b","Type":"ContainerStarted","Data":"fd5913a59424afae08dbe8fcb4fe7e4a6df73fc8236be1dae26a2e9abff9a3d9"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.711035 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-v4hbg" podStartSLOduration=130.711002342 podStartE2EDuration="2m10.711002342s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:04.709869461 +0000 UTC m=+157.080979872" watchObservedRunningTime="2025-12-01 17:28:04.711002342 +0000 UTC m=+157.082112753" Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.713998 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m8zhv" podStartSLOduration=131.713988674 podStartE2EDuration="2m11.713988674s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:03.722636637 +0000 UTC m=+156.093747048" watchObservedRunningTime="2025-12-01 17:28:04.713988674 +0000 UTC m=+157.085099085" Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.730077 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-v4fcn" event={"ID":"b0f5f4f9-02b9-48f1-918e-985b836626ad","Type":"ContainerStarted","Data":"e863b6b744923589766f6e9e8fa4b96149a81d7fa53f160152437f3fe0ab64d9"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.732403 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" event={"ID":"da5b10e6-e695-4e8a-84e5-329870d5736e","Type":"ContainerStarted","Data":"f0f40a4b58c7069c6b60cbcffa6cae4f0ad20515f0970b7fe96fe75d62d685d8"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.735342 4868 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wnnm4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.735383 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" podUID="da5b10e6-e695-4e8a-84e5-329870d5736e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.736357 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:04 crc kubenswrapper[4868]: E1201 17:28:04.737355 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:05.23734145 +0000 UTC m=+157.608451861 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.738296 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vw9z9" event={"ID":"2bf15be5-1380-4b58-934d-7793e22541b4","Type":"ContainerStarted","Data":"69e3bc76be6c52d3cf1a3b49daa123b5a305283bb3717d56a7bae176f947e7f7"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.739808 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-v8wh9" event={"ID":"3bfca371-4e3e-4373-9589-1e1ee7042f85","Type":"ContainerStarted","Data":"5e3544cc7f3d8ea43b8d1c1278072bcdf635ab55a719a5ea51c456924e08e5b7"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.740875 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk" event={"ID":"95e7a05d-5fa9-47ce-8cf3-d261fea05465","Type":"ContainerStarted","Data":"d2e1ffe4efa4afe9d919365591d2e56f2b432850c1bb5bcd4e9c57088287ee37"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.741263 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk" Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.743587 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-42p77" event={"ID":"7ca45004-4f76-480b-b8b8-68178408dde2","Type":"ContainerStarted","Data":"84c5be23654bb8f00488a9e8aa8709a473070d7f28d0425996769a78cafa3d85"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.743605 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-42p77" event={"ID":"7ca45004-4f76-480b-b8b8-68178408dde2","Type":"ContainerStarted","Data":"a70ebd9797353b8c449490c284c52f5526ea1208e1be941359960acd1d513ede"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.760748 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-q58ld" event={"ID":"a5bca2c5-996d-49a5-92ac-1d38c96134eb","Type":"ContainerStarted","Data":"a470cd0893bd4e62b196fec949d1e9504e70957ce114b86088defca97a024f9b"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.785746 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fgjv4" event={"ID":"9a983c06-7704-4c22-8c1b-d0cfcdc67ee8","Type":"ContainerStarted","Data":"5b239c61624c44ea02174a6cb75056a1cea7ca8dceef7d56652759d75a5b0a36"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.798638 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" event={"ID":"71ff3201-293e-4a9e-ae2a-006c945714ef","Type":"ContainerStarted","Data":"41b00fd38383ed24bd70762e60394d2bc1f833baaf2e0457576d7a44e3035085"} Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.799754 4868 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4hvgn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.800786 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" podUID="22cc1879-5e5e-4a06-ba88-f92033382b90" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.800918 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.799916 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.801150 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.814759 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" podStartSLOduration=131.814723407 podStartE2EDuration="2m11.814723407s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:04.737514265 +0000 UTC m=+157.108624676" watchObservedRunningTime="2025-12-01 17:28:04.814723407 +0000 UTC m=+157.185833818" Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.815224 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k7ghg" podStartSLOduration=130.815218011 podStartE2EDuration="2m10.815218011s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:04.797676786 +0000 UTC m=+157.168787197" watchObservedRunningTime="2025-12-01 17:28:04.815218011 +0000 UTC m=+157.186328422" Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.833914 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mt2zl" Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.839682 4868 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-7q4rx container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.839748 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" podUID="71ff3201-293e-4a9e-ae2a-006c945714ef" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.843858 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:04 crc kubenswrapper[4868]: E1201 17:28:04.859789 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:05.359764822 +0000 UTC m=+157.730875233 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.860514 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-c4nzd" podStartSLOduration=130.860492032 podStartE2EDuration="2m10.860492032s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:04.858906178 +0000 UTC m=+157.230016599" watchObservedRunningTime="2025-12-01 17:28:04.860492032 +0000 UTC m=+157.231602443" Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.933428 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5dqtb" podStartSLOduration=130.933401557 podStartE2EDuration="2m10.933401557s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:04.930742993 +0000 UTC m=+157.301853414" watchObservedRunningTime="2025-12-01 17:28:04.933401557 +0000 UTC m=+157.304511958" Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.945662 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:04 crc kubenswrapper[4868]: E1201 17:28:04.946072 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:05.446052386 +0000 UTC m=+157.817162797 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:04 crc kubenswrapper[4868]: I1201 17:28:04.962887 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-42p77" podStartSLOduration=131.96286113 podStartE2EDuration="2m11.96286113s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:04.961416001 +0000 UTC m=+157.332526412" watchObservedRunningTime="2025-12-01 17:28:04.96286113 +0000 UTC m=+157.333971541" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.047388 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:05 crc kubenswrapper[4868]: E1201 17:28:05.047842 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:05.547823417 +0000 UTC m=+157.918933818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.049669 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-q58ld" podStartSLOduration=131.049643317 podStartE2EDuration="2m11.049643317s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:04.997197068 +0000 UTC m=+157.368307479" watchObservedRunningTime="2025-12-01 17:28:05.049643317 +0000 UTC m=+157.420753728" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.049817 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vw9z9" podStartSLOduration=131.049812482 podStartE2EDuration="2m11.049812482s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:05.045589166 +0000 UTC m=+157.416699577" watchObservedRunningTime="2025-12-01 17:28:05.049812482 +0000 UTC m=+157.420922893" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.117774 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk" podStartSLOduration=131.117753529 podStartE2EDuration="2m11.117753529s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:05.115020453 +0000 UTC m=+157.486130864" watchObservedRunningTime="2025-12-01 17:28:05.117753529 +0000 UTC m=+157.488863940" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.163027 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:05 crc kubenswrapper[4868]: E1201 17:28:05.163274 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:05.663252557 +0000 UTC m=+158.034362968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.185390 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-v4fcn" podStartSLOduration=10.185372198 podStartE2EDuration="10.185372198s" podCreationTimestamp="2025-12-01 17:27:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:05.147670506 +0000 UTC m=+157.518780917" watchObservedRunningTime="2025-12-01 17:28:05.185372198 +0000 UTC m=+157.556482609" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.264831 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:05 crc kubenswrapper[4868]: E1201 17:28:05.265308 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:05.765286625 +0000 UTC m=+158.136397036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.365832 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:05 crc kubenswrapper[4868]: E1201 17:28:05.366184 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:05.866126231 +0000 UTC m=+158.237236642 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.416538 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.446778 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-hwgjb" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.468499 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:05 crc kubenswrapper[4868]: E1201 17:28:05.471344 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:05.971324768 +0000 UTC m=+158.342435179 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.486091 4868 patch_prober.go:28] interesting pod/router-default-5444994796-bddbd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 17:28:05 crc kubenswrapper[4868]: [-]has-synced failed: reason withheld Dec 01 17:28:05 crc kubenswrapper[4868]: [+]process-running ok Dec 01 17:28:05 crc kubenswrapper[4868]: healthz check failed Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.486194 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bddbd" podUID="886b276a-b6e7-4ce1-b6cf-dc269d9180bb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.486211 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" podStartSLOduration=131.486190318 podStartE2EDuration="2m11.486190318s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:05.187105845 +0000 UTC m=+157.558216256" watchObservedRunningTime="2025-12-01 17:28:05.486190318 +0000 UTC m=+157.857300729" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.577597 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:05 crc kubenswrapper[4868]: E1201 17:28:05.577762 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:06.077720107 +0000 UTC m=+158.448830518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.577841 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:05 crc kubenswrapper[4868]: E1201 17:28:05.578228 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:06.078221711 +0000 UTC m=+158.449332122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.678849 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:05 crc kubenswrapper[4868]: E1201 17:28:05.679113 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:06.179071107 +0000 UTC m=+158.550181518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.782829 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:05 crc kubenswrapper[4868]: E1201 17:28:05.783276 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:06.283259095 +0000 UTC m=+158.654369506 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.844355 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" event={"ID":"d81ae1af-29ea-4b7f-a935-1faa8973fcc9","Type":"ContainerStarted","Data":"35e266d3cda59fd59d14a5d2cfb933f33bed1f42c3c11e838e8bacc05ae8ef46"} Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.847624 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fgjv4" event={"ID":"9a983c06-7704-4c22-8c1b-d0cfcdc67ee8","Type":"ContainerStarted","Data":"4ad94457c18665a6f814f4db767c6e0a803c65b7fad3124732f561919c8572b6"} Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.854484 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-v8wh9" event={"ID":"3bfca371-4e3e-4373-9589-1e1ee7042f85","Type":"ContainerStarted","Data":"22126fb1de8aac69d7c8417d97dae1366e22e65a2847f8a947caf009bc2d587d"} Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.867922 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-nzh9r" event={"ID":"7b50f7b4-dd55-4255-9b68-a942ff4a086e","Type":"ContainerStarted","Data":"dbc189e50c254474e09202e7e2a212bd06c5235f47d561d4405e23485f7c7f75"} Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.868461 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-nzh9r" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.871283 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.871347 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.876189 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b89j6"] Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.877206 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.884168 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:05 crc kubenswrapper[4868]: E1201 17:28:05.884277 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:06.384259135 +0000 UTC m=+158.755369546 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.885979 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.886051 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/262076bc-ef10-4531-a6f9-f32148ee7a3e-utilities\") pod \"certified-operators-b89j6\" (UID: \"262076bc-ef10-4531-a6f9-f32148ee7a3e\") " pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.886165 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7qrp\" (UniqueName: \"kubernetes.io/projected/262076bc-ef10-4531-a6f9-f32148ee7a3e-kube-api-access-n7qrp\") pod \"certified-operators-b89j6\" (UID: \"262076bc-ef10-4531-a6f9-f32148ee7a3e\") " pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.886344 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/262076bc-ef10-4531-a6f9-f32148ee7a3e-catalog-content\") pod \"certified-operators-b89j6\" (UID: \"262076bc-ef10-4531-a6f9-f32148ee7a3e\") " pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:28:05 crc kubenswrapper[4868]: E1201 17:28:05.888134 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:06.388122622 +0000 UTC m=+158.759233033 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.896526 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7q4rx" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.908468 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.910826 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b89j6"] Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.941380 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xtcsm"] Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.942430 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.950660 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.989373 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.989640 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/262076bc-ef10-4531-a6f9-f32148ee7a3e-utilities\") pod \"certified-operators-b89j6\" (UID: \"262076bc-ef10-4531-a6f9-f32148ee7a3e\") " pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.989678 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7qrp\" (UniqueName: \"kubernetes.io/projected/262076bc-ef10-4531-a6f9-f32148ee7a3e-kube-api-access-n7qrp\") pod \"certified-operators-b89j6\" (UID: \"262076bc-ef10-4531-a6f9-f32148ee7a3e\") " pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.989711 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/262076bc-ef10-4531-a6f9-f32148ee7a3e-catalog-content\") pod \"certified-operators-b89j6\" (UID: \"262076bc-ef10-4531-a6f9-f32148ee7a3e\") " pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.990235 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/262076bc-ef10-4531-a6f9-f32148ee7a3e-catalog-content\") pod \"certified-operators-b89j6\" (UID: \"262076bc-ef10-4531-a6f9-f32148ee7a3e\") " pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:28:05 crc kubenswrapper[4868]: E1201 17:28:05.990678 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:06.490658425 +0000 UTC m=+158.861768836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:05 crc kubenswrapper[4868]: I1201 17:28:05.990916 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/262076bc-ef10-4531-a6f9-f32148ee7a3e-utilities\") pod \"certified-operators-b89j6\" (UID: \"262076bc-ef10-4531-a6f9-f32148ee7a3e\") " pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.057085 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-fgjv4" podStartSLOduration=132.057045409 podStartE2EDuration="2m12.057045409s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:05.988610028 +0000 UTC m=+158.359720439" watchObservedRunningTime="2025-12-01 17:28:06.057045409 +0000 UTC m=+158.428155820" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.061932 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-v8wh9" podStartSLOduration=133.061916433 podStartE2EDuration="2m13.061916433s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:06.051463775 +0000 UTC m=+158.422574186" watchObservedRunningTime="2025-12-01 17:28:06.061916433 +0000 UTC m=+158.433026844" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.062707 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xtcsm"] Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.091071 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-catalog-content\") pod \"community-operators-xtcsm\" (UID: \"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06\") " pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.091172 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvxwm\" (UniqueName: \"kubernetes.io/projected/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-kube-api-access-pvxwm\") pod \"community-operators-xtcsm\" (UID: \"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06\") " pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.091212 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-utilities\") pod \"community-operators-xtcsm\" (UID: \"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06\") " pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.091256 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:06 crc kubenswrapper[4868]: E1201 17:28:06.091631 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:06.591614634 +0000 UTC m=+158.962725045 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.093020 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7qrp\" (UniqueName: \"kubernetes.io/projected/262076bc-ef10-4531-a6f9-f32148ee7a3e-kube-api-access-n7qrp\") pod \"certified-operators-b89j6\" (UID: \"262076bc-ef10-4531-a6f9-f32148ee7a3e\") " pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.194620 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.194909 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-catalog-content\") pod \"community-operators-xtcsm\" (UID: \"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06\") " pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.195006 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvxwm\" (UniqueName: \"kubernetes.io/projected/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-kube-api-access-pvxwm\") pod \"community-operators-xtcsm\" (UID: \"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06\") " pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.195031 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-utilities\") pod \"community-operators-xtcsm\" (UID: \"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06\") " pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.195571 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-utilities\") pod \"community-operators-xtcsm\" (UID: \"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06\") " pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:28:06 crc kubenswrapper[4868]: E1201 17:28:06.195684 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:06.695664828 +0000 UTC m=+159.066775239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.195898 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-catalog-content\") pod \"community-operators-xtcsm\" (UID: \"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06\") " pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.210799 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.261665 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c9tpp"] Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.263085 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.306910 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.307023 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b8mw\" (UniqueName: \"kubernetes.io/projected/b2c7b9b4-a824-4639-9dfa-21778b900c9a-kube-api-access-4b8mw\") pod \"certified-operators-c9tpp\" (UID: \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\") " pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.307059 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2c7b9b4-a824-4639-9dfa-21778b900c9a-utilities\") pod \"certified-operators-c9tpp\" (UID: \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\") " pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.307084 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2c7b9b4-a824-4639-9dfa-21778b900c9a-catalog-content\") pod \"certified-operators-c9tpp\" (UID: \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\") " pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:28:06 crc kubenswrapper[4868]: E1201 17:28:06.307436 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:06.807424726 +0000 UTC m=+159.178535137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.357439 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvxwm\" (UniqueName: \"kubernetes.io/projected/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-kube-api-access-pvxwm\") pod \"community-operators-xtcsm\" (UID: \"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06\") " pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.402648 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-nzh9r" podStartSLOduration=11.402628036 podStartE2EDuration="11.402628036s" podCreationTimestamp="2025-12-01 17:27:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:06.312716702 +0000 UTC m=+158.683827123" watchObservedRunningTime="2025-12-01 17:28:06.402628036 +0000 UTC m=+158.773738447" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.457392 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.457978 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b8mw\" (UniqueName: \"kubernetes.io/projected/b2c7b9b4-a824-4639-9dfa-21778b900c9a-kube-api-access-4b8mw\") pod \"certified-operators-c9tpp\" (UID: \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\") " pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.458020 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2c7b9b4-a824-4639-9dfa-21778b900c9a-utilities\") pod \"certified-operators-c9tpp\" (UID: \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\") " pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.458050 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2c7b9b4-a824-4639-9dfa-21778b900c9a-catalog-content\") pod \"certified-operators-c9tpp\" (UID: \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\") " pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.458553 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2c7b9b4-a824-4639-9dfa-21778b900c9a-catalog-content\") pod \"certified-operators-c9tpp\" (UID: \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\") " pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:28:06 crc kubenswrapper[4868]: E1201 17:28:06.459453 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:06.959433905 +0000 UTC m=+159.330544316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.460110 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2c7b9b4-a824-4639-9dfa-21778b900c9a-utilities\") pod \"certified-operators-c9tpp\" (UID: \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\") " pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.461084 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c9tpp"] Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.477040 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hz4pc"] Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.479168 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.517814 4868 patch_prober.go:28] interesting pod/router-default-5444994796-bddbd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 17:28:06 crc kubenswrapper[4868]: [-]has-synced failed: reason withheld Dec 01 17:28:06 crc kubenswrapper[4868]: [+]process-running ok Dec 01 17:28:06 crc kubenswrapper[4868]: healthz check failed Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.517894 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bddbd" podUID="886b276a-b6e7-4ce1-b6cf-dc269d9180bb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.519008 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b8mw\" (UniqueName: \"kubernetes.io/projected/b2c7b9b4-a824-4639-9dfa-21778b900c9a-kube-api-access-4b8mw\") pod \"certified-operators-c9tpp\" (UID: \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\") " pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.551014 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hz4pc"] Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.584144 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-catalog-content\") pod \"community-operators-hz4pc\" (UID: \"d7c0aa99-172b-41b7-80e5-2cbc6532cae1\") " pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.584217 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-utilities\") pod \"community-operators-hz4pc\" (UID: \"d7c0aa99-172b-41b7-80e5-2cbc6532cae1\") " pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.584281 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2jhx\" (UniqueName: \"kubernetes.io/projected/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-kube-api-access-n2jhx\") pod \"community-operators-hz4pc\" (UID: \"d7c0aa99-172b-41b7-80e5-2cbc6532cae1\") " pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.584312 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:06 crc kubenswrapper[4868]: E1201 17:28:06.584670 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:07.084653385 +0000 UTC m=+159.455763796 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.612342 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.634105 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnnm4" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.649809 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.686405 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.686614 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2jhx\" (UniqueName: \"kubernetes.io/projected/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-kube-api-access-n2jhx\") pod \"community-operators-hz4pc\" (UID: \"d7c0aa99-172b-41b7-80e5-2cbc6532cae1\") " pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.686679 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-catalog-content\") pod \"community-operators-hz4pc\" (UID: \"d7c0aa99-172b-41b7-80e5-2cbc6532cae1\") " pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.686720 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-utilities\") pod \"community-operators-hz4pc\" (UID: \"d7c0aa99-172b-41b7-80e5-2cbc6532cae1\") " pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.687256 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-utilities\") pod \"community-operators-hz4pc\" (UID: \"d7c0aa99-172b-41b7-80e5-2cbc6532cae1\") " pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:28:06 crc kubenswrapper[4868]: E1201 17:28:06.689332 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:07.189307586 +0000 UTC m=+159.560417997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.705932 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-catalog-content\") pod \"community-operators-hz4pc\" (UID: \"d7c0aa99-172b-41b7-80e5-2cbc6532cae1\") " pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.739959 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2jhx\" (UniqueName: \"kubernetes.io/projected/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-kube-api-access-n2jhx\") pod \"community-operators-hz4pc\" (UID: \"d7c0aa99-172b-41b7-80e5-2cbc6532cae1\") " pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.788197 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:06 crc kubenswrapper[4868]: E1201 17:28:06.788679 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:07.288664341 +0000 UTC m=+159.659774762 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.892345 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:06 crc kubenswrapper[4868]: E1201 17:28:06.892778 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:07.392759097 +0000 UTC m=+159.763869508 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.897329 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:28:06 crc kubenswrapper[4868]: I1201 17:28:06.919005 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" event={"ID":"d81ae1af-29ea-4b7f-a935-1faa8973fcc9","Type":"ContainerStarted","Data":"c9e421c94a252fb17776fbaa4ce97f418d656470002898ee5cf38d3ef867783c"} Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.007352 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:07 crc kubenswrapper[4868]: E1201 17:28:07.009820 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:07.50980141 +0000 UTC m=+159.880911821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.085164 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b89j6"] Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.109696 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:07 crc kubenswrapper[4868]: E1201 17:28:07.110025 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:07.609990109 +0000 UTC m=+159.981100520 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.110085 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:07 crc kubenswrapper[4868]: E1201 17:28:07.110475 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:07.610467332 +0000 UTC m=+159.981577743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.213584 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:07 crc kubenswrapper[4868]: E1201 17:28:07.214132 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:07.714111615 +0000 UTC m=+160.085222026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.250587 4868 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.316005 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:07 crc kubenswrapper[4868]: E1201 17:28:07.316520 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 17:28:07.816500363 +0000 UTC m=+160.187610774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-x5gxr" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.397314 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xtcsm"] Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.417866 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:07 crc kubenswrapper[4868]: E1201 17:28:07.418968 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 17:28:07.918908073 +0000 UTC m=+160.290018474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.475255 4868 patch_prober.go:28] interesting pod/router-default-5444994796-bddbd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 17:28:07 crc kubenswrapper[4868]: [-]has-synced failed: reason withheld Dec 01 17:28:07 crc kubenswrapper[4868]: [+]process-running ok Dec 01 17:28:07 crc kubenswrapper[4868]: healthz check failed Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.476520 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bddbd" podUID="886b276a-b6e7-4ce1-b6cf-dc269d9180bb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.491312 4868 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-01T17:28:07.250618933Z","Handler":null,"Name":""} Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.498411 4868 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.498450 4868 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.520544 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.527643 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-262xt" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.529237 4868 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.529310 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.669926 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c9tpp"] Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.690250 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hz4pc"] Dec 01 17:28:07 crc kubenswrapper[4868]: W1201 17:28:07.690994 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2c7b9b4_a824_4639_9dfa_21778b900c9a.slice/crio-553e5dadc5b0e3d510394c44df2681b51d2f9ffc58f071004dd3aedea85f5c66 WatchSource:0}: Error finding container 553e5dadc5b0e3d510394c44df2681b51d2f9ffc58f071004dd3aedea85f5c66: Status 404 returned error can't find the container with id 553e5dadc5b0e3d510394c44df2681b51d2f9ffc58f071004dd3aedea85f5c66 Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.710616 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fnrd7"] Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.713460 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.716702 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.719867 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-x5gxr\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:07 crc kubenswrapper[4868]: W1201 17:28:07.757315 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7c0aa99_172b_41b7_80e5_2cbc6532cae1.slice/crio-f229bd27517b9ad6dde6edb651199086ad6f28f5e84ea204572f36d363476cb2 WatchSource:0}: Error finding container f229bd27517b9ad6dde6edb651199086ad6f28f5e84ea204572f36d363476cb2: Status 404 returned error can't find the container with id f229bd27517b9ad6dde6edb651199086ad6f28f5e84ea204572f36d363476cb2 Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.825461 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnrd7"] Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.837651 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.838052 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c64j5\" (UniqueName: \"kubernetes.io/projected/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-kube-api-access-c64j5\") pod \"redhat-marketplace-fnrd7\" (UID: \"de9baf7b-7da5-4cd3-908a-d3bc26f918f1\") " pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.838096 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-utilities\") pod \"redhat-marketplace-fnrd7\" (UID: \"de9baf7b-7da5-4cd3-908a-d3bc26f918f1\") " pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.838193 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-catalog-content\") pod \"redhat-marketplace-fnrd7\" (UID: \"de9baf7b-7da5-4cd3-908a-d3bc26f918f1\") " pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.885337 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.887240 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.887809 4868 patch_prober.go:28] interesting pod/console-f9d7485db-c7pwq container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.887874 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-c7pwq" podUID="e90813d0-8960-450f-aaa4-047341c50947" containerName="console" probeResult="failure" output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.908026 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.928002 4868 generic.go:334] "Generic (PLEG): container finished" podID="962e4f14-b280-4ffa-bde0-6cc4b0d86144" containerID="83cbd781a24db21510be592699a361e90e624e963509bd92bbef568a0b050db1" exitCode=0 Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.928094 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" event={"ID":"962e4f14-b280-4ffa-bde0-6cc4b0d86144","Type":"ContainerDied","Data":"83cbd781a24db21510be592699a361e90e624e963509bd92bbef568a0b050db1"} Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.929861 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9tpp" event={"ID":"b2c7b9b4-a824-4639-9dfa-21778b900c9a","Type":"ContainerStarted","Data":"553e5dadc5b0e3d510394c44df2681b51d2f9ffc58f071004dd3aedea85f5c66"} Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.935184 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hz4pc" event={"ID":"d7c0aa99-172b-41b7-80e5-2cbc6532cae1","Type":"ContainerStarted","Data":"f229bd27517b9ad6dde6edb651199086ad6f28f5e84ea204572f36d363476cb2"} Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.939435 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-catalog-content\") pod \"redhat-marketplace-fnrd7\" (UID: \"de9baf7b-7da5-4cd3-908a-d3bc26f918f1\") " pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.939490 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c64j5\" (UniqueName: \"kubernetes.io/projected/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-kube-api-access-c64j5\") pod \"redhat-marketplace-fnrd7\" (UID: \"de9baf7b-7da5-4cd3-908a-d3bc26f918f1\") " pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.939513 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-utilities\") pod \"redhat-marketplace-fnrd7\" (UID: \"de9baf7b-7da5-4cd3-908a-d3bc26f918f1\") " pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.940477 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-catalog-content\") pod \"redhat-marketplace-fnrd7\" (UID: \"de9baf7b-7da5-4cd3-908a-d3bc26f918f1\") " pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.941215 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-utilities\") pod \"redhat-marketplace-fnrd7\" (UID: \"de9baf7b-7da5-4cd3-908a-d3bc26f918f1\") " pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.954068 4868 generic.go:334] "Generic (PLEG): container finished" podID="262076bc-ef10-4531-a6f9-f32148ee7a3e" containerID="df09d57f0923446e8127960676005547523e9e1115061159df1ebd582de85c82" exitCode=0 Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.954243 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b89j6" event={"ID":"262076bc-ef10-4531-a6f9-f32148ee7a3e","Type":"ContainerDied","Data":"df09d57f0923446e8127960676005547523e9e1115061159df1ebd582de85c82"} Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.954310 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b89j6" event={"ID":"262076bc-ef10-4531-a6f9-f32148ee7a3e","Type":"ContainerStarted","Data":"13f130cd42622905a02755a1d0729f1ea252cfeaf58d815d802363eab0b52908"} Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.959312 4868 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.973406 4868 generic.go:334] "Generic (PLEG): container finished" podID="b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" containerID="afe3497f1a9177edd7026915922c659b3dc467b09e9a055b2765bcdd7fdae125" exitCode=0 Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.973508 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xtcsm" event={"ID":"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06","Type":"ContainerDied","Data":"afe3497f1a9177edd7026915922c659b3dc467b09e9a055b2765bcdd7fdae125"} Dec 01 17:28:07 crc kubenswrapper[4868]: I1201 17:28:07.973541 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xtcsm" event={"ID":"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06","Type":"ContainerStarted","Data":"a4dc3c532e7ba228c311a079eb860960773fe3fddb0b5e782ea0af8da69a4c7b"} Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.013819 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" event={"ID":"d81ae1af-29ea-4b7f-a935-1faa8973fcc9","Type":"ContainerStarted","Data":"517975f92310abb74efb3a868720135e6a1ed77f2c287835e0bbeb869fa38316"} Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.013903 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" event={"ID":"d81ae1af-29ea-4b7f-a935-1faa8973fcc9","Type":"ContainerStarted","Data":"52685447d06f9b24e408dbd8836904ff4234ade2825dc6813f207e7b24519c56"} Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.014343 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.072594 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c64j5\" (UniqueName: \"kubernetes.io/projected/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-kube-api-access-c64j5\") pod \"redhat-marketplace-fnrd7\" (UID: \"de9baf7b-7da5-4cd3-908a-d3bc26f918f1\") " pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.082041 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.082116 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.082224 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.082267 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.128866 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.139316 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.224574 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.470665 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.484238 4868 patch_prober.go:28] interesting pod/router-default-5444994796-bddbd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 17:28:08 crc kubenswrapper[4868]: [-]has-synced failed: reason withheld Dec 01 17:28:08 crc kubenswrapper[4868]: [+]process-running ok Dec 01 17:28:08 crc kubenswrapper[4868]: healthz check failed Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.484304 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bddbd" podUID="886b276a-b6e7-4ce1-b6cf-dc269d9180bb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.502288 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4tjsn"] Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.503370 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.570127 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-utilities\") pod \"redhat-marketplace-4tjsn\" (UID: \"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef\") " pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.570184 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlbmh\" (UniqueName: \"kubernetes.io/projected/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-kube-api-access-zlbmh\") pod \"redhat-marketplace-4tjsn\" (UID: \"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef\") " pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.570235 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-catalog-content\") pod \"redhat-marketplace-4tjsn\" (UID: \"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef\") " pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.634638 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tjsn"] Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.672720 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlbmh\" (UniqueName: \"kubernetes.io/projected/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-kube-api-access-zlbmh\") pod \"redhat-marketplace-4tjsn\" (UID: \"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef\") " pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.673234 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-catalog-content\") pod \"redhat-marketplace-4tjsn\" (UID: \"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef\") " pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.673312 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-utilities\") pod \"redhat-marketplace-4tjsn\" (UID: \"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef\") " pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.673692 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-utilities\") pod \"redhat-marketplace-4tjsn\" (UID: \"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef\") " pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.674588 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-catalog-content\") pod \"redhat-marketplace-4tjsn\" (UID: \"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef\") " pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.744831 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlbmh\" (UniqueName: \"kubernetes.io/projected/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-kube-api-access-zlbmh\") pod \"redhat-marketplace-4tjsn\" (UID: \"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef\") " pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.840695 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-g8kh5" podStartSLOduration=13.840666871 podStartE2EDuration="13.840666871s" podCreationTimestamp="2025-12-01 17:27:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:08.83810605 +0000 UTC m=+161.209216461" watchObservedRunningTime="2025-12-01 17:28:08.840666871 +0000 UTC m=+161.211777282" Dec 01 17:28:08 crc kubenswrapper[4868]: I1201 17:28:08.969303 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.017155 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnrd7"] Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.065779 4868 generic.go:334] "Generic (PLEG): container finished" podID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" containerID="cb9d980c782a5777e72b3ce43d7000fe2a846d22a628e55ea15ad130389c8cd6" exitCode=0 Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.066257 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9tpp" event={"ID":"b2c7b9b4-a824-4639-9dfa-21778b900c9a","Type":"ContainerDied","Data":"cb9d980c782a5777e72b3ce43d7000fe2a846d22a628e55ea15ad130389c8cd6"} Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.100863 4868 generic.go:334] "Generic (PLEG): container finished" podID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" containerID="959f721f35672a4e9bd5929a14dbede468422b3668484db48720077c48ce132e" exitCode=0 Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.101701 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hz4pc" event={"ID":"d7c0aa99-172b-41b7-80e5-2cbc6532cae1","Type":"ContainerDied","Data":"959f721f35672a4e9bd5929a14dbede468422b3668484db48720077c48ce132e"} Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.119924 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2f8vp"] Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.133734 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.136176 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.178502 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2f8vp"] Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.200172 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-x5gxr"] Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.297542 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwg6w\" (UniqueName: \"kubernetes.io/projected/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-kube-api-access-jwg6w\") pod \"redhat-operators-2f8vp\" (UID: \"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e\") " pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.297905 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-catalog-content\") pod \"redhat-operators-2f8vp\" (UID: \"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e\") " pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.297979 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-utilities\") pod \"redhat-operators-2f8vp\" (UID: \"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e\") " pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.403142 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwg6w\" (UniqueName: \"kubernetes.io/projected/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-kube-api-access-jwg6w\") pod \"redhat-operators-2f8vp\" (UID: \"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e\") " pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.403628 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-catalog-content\") pod \"redhat-operators-2f8vp\" (UID: \"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e\") " pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.403697 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-utilities\") pod \"redhat-operators-2f8vp\" (UID: \"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e\") " pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.405199 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-utilities\") pod \"redhat-operators-2f8vp\" (UID: \"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e\") " pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.405721 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-catalog-content\") pod \"redhat-operators-2f8vp\" (UID: \"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e\") " pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.428102 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.431231 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.439769 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwg6w\" (UniqueName: \"kubernetes.io/projected/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-kube-api-access-jwg6w\") pod \"redhat-operators-2f8vp\" (UID: \"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e\") " pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.440091 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.440369 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.449224 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.499686 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.504976 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.515490 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-bddbd" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.519023 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pwdv6"] Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.520502 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.526112 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pwdv6"] Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.596085 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tjsn"] Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.606739 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/77bde072-b3d6-494f-88e4-0732ec5f3ebb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"77bde072-b3d6-494f-88e4-0732ec5f3ebb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.606804 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/77bde072-b3d6-494f-88e4-0732ec5f3ebb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"77bde072-b3d6-494f-88e4-0732ec5f3ebb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.613258 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.707735 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6332f65b-5f50-4e61-9f2d-b4333fda483b-utilities\") pod \"redhat-operators-pwdv6\" (UID: \"6332f65b-5f50-4e61-9f2d-b4333fda483b\") " pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.708056 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgls8\" (UniqueName: \"kubernetes.io/projected/6332f65b-5f50-4e61-9f2d-b4333fda483b-kube-api-access-pgls8\") pod \"redhat-operators-pwdv6\" (UID: \"6332f65b-5f50-4e61-9f2d-b4333fda483b\") " pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.708088 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/77bde072-b3d6-494f-88e4-0732ec5f3ebb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"77bde072-b3d6-494f-88e4-0732ec5f3ebb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.708108 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6332f65b-5f50-4e61-9f2d-b4333fda483b-catalog-content\") pod \"redhat-operators-pwdv6\" (UID: \"6332f65b-5f50-4e61-9f2d-b4333fda483b\") " pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.708130 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/77bde072-b3d6-494f-88e4-0732ec5f3ebb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"77bde072-b3d6-494f-88e4-0732ec5f3ebb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.708569 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/77bde072-b3d6-494f-88e4-0732ec5f3ebb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"77bde072-b3d6-494f-88e4-0732ec5f3ebb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 17:28:09 crc kubenswrapper[4868]: E1201 17:28:09.772477 4868 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde9baf7b_7da5_4cd3_908a_d3bc26f918f1.slice/crio-conmon-87b15060fb50c05e6ee2ee304b4c43aaf0c6591ab275f4561e0afb8d32be4b8c.scope\": RecentStats: unable to find data in memory cache]" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.778003 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/77bde072-b3d6-494f-88e4-0732ec5f3ebb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"77bde072-b3d6-494f-88e4-0732ec5f3ebb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.808796 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nm6hg\" (UniqueName: \"kubernetes.io/projected/962e4f14-b280-4ffa-bde0-6cc4b0d86144-kube-api-access-nm6hg\") pod \"962e4f14-b280-4ffa-bde0-6cc4b0d86144\" (UID: \"962e4f14-b280-4ffa-bde0-6cc4b0d86144\") " Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.808884 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/962e4f14-b280-4ffa-bde0-6cc4b0d86144-secret-volume\") pod \"962e4f14-b280-4ffa-bde0-6cc4b0d86144\" (UID: \"962e4f14-b280-4ffa-bde0-6cc4b0d86144\") " Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.809010 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/962e4f14-b280-4ffa-bde0-6cc4b0d86144-config-volume\") pod \"962e4f14-b280-4ffa-bde0-6cc4b0d86144\" (UID: \"962e4f14-b280-4ffa-bde0-6cc4b0d86144\") " Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.809344 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6332f65b-5f50-4e61-9f2d-b4333fda483b-utilities\") pod \"redhat-operators-pwdv6\" (UID: \"6332f65b-5f50-4e61-9f2d-b4333fda483b\") " pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.809380 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgls8\" (UniqueName: \"kubernetes.io/projected/6332f65b-5f50-4e61-9f2d-b4333fda483b-kube-api-access-pgls8\") pod \"redhat-operators-pwdv6\" (UID: \"6332f65b-5f50-4e61-9f2d-b4333fda483b\") " pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.809425 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6332f65b-5f50-4e61-9f2d-b4333fda483b-catalog-content\") pod \"redhat-operators-pwdv6\" (UID: \"6332f65b-5f50-4e61-9f2d-b4333fda483b\") " pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.810184 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6332f65b-5f50-4e61-9f2d-b4333fda483b-catalog-content\") pod \"redhat-operators-pwdv6\" (UID: \"6332f65b-5f50-4e61-9f2d-b4333fda483b\") " pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.812998 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6332f65b-5f50-4e61-9f2d-b4333fda483b-utilities\") pod \"redhat-operators-pwdv6\" (UID: \"6332f65b-5f50-4e61-9f2d-b4333fda483b\") " pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.813702 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/962e4f14-b280-4ffa-bde0-6cc4b0d86144-config-volume" (OuterVolumeSpecName: "config-volume") pod "962e4f14-b280-4ffa-bde0-6cc4b0d86144" (UID: "962e4f14-b280-4ffa-bde0-6cc4b0d86144"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.829460 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.836217 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/962e4f14-b280-4ffa-bde0-6cc4b0d86144-kube-api-access-nm6hg" (OuterVolumeSpecName: "kube-api-access-nm6hg") pod "962e4f14-b280-4ffa-bde0-6cc4b0d86144" (UID: "962e4f14-b280-4ffa-bde0-6cc4b0d86144"). InnerVolumeSpecName "kube-api-access-nm6hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.837105 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962e4f14-b280-4ffa-bde0-6cc4b0d86144-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "962e4f14-b280-4ffa-bde0-6cc4b0d86144" (UID: "962e4f14-b280-4ffa-bde0-6cc4b0d86144"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.847192 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgls8\" (UniqueName: \"kubernetes.io/projected/6332f65b-5f50-4e61-9f2d-b4333fda483b-kube-api-access-pgls8\") pod \"redhat-operators-pwdv6\" (UID: \"6332f65b-5f50-4e61-9f2d-b4333fda483b\") " pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.852890 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.914027 4868 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/962e4f14-b280-4ffa-bde0-6cc4b0d86144-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.914451 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nm6hg\" (UniqueName: \"kubernetes.io/projected/962e4f14-b280-4ffa-bde0-6cc4b0d86144-kube-api-access-nm6hg\") on node \"crc\" DevicePath \"\"" Dec 01 17:28:09 crc kubenswrapper[4868]: I1201 17:28:09.914463 4868 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/962e4f14-b280-4ffa-bde0-6cc4b0d86144-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.169107 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2f8vp"] Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.187850 4868 generic.go:334] "Generic (PLEG): container finished" podID="de9baf7b-7da5-4cd3-908a-d3bc26f918f1" containerID="87b15060fb50c05e6ee2ee304b4c43aaf0c6591ab275f4561e0afb8d32be4b8c" exitCode=0 Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.262416 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.270773 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.276104 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 17:28:10 crc kubenswrapper[4868]: E1201 17:28:10.276538 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="962e4f14-b280-4ffa-bde0-6cc4b0d86144" containerName="collect-profiles" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.276562 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="962e4f14-b280-4ffa-bde0-6cc4b0d86144" containerName="collect-profiles" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.276698 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="962e4f14-b280-4ffa-bde0-6cc4b0d86144" containerName="collect-profiles" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.277187 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnrd7" event={"ID":"de9baf7b-7da5-4cd3-908a-d3bc26f918f1","Type":"ContainerDied","Data":"87b15060fb50c05e6ee2ee304b4c43aaf0c6591ab275f4561e0afb8d32be4b8c"} Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.277236 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnrd7" event={"ID":"de9baf7b-7da5-4cd3-908a-d3bc26f918f1","Type":"ContainerStarted","Data":"0b1ffba19d22107b3ab1498e7c1358b95d9dd1fd97f8609949d873f9f3a2ff05"} Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.277251 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" event={"ID":"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6","Type":"ContainerStarted","Data":"ac6ebb21f802c3122ef3b0fcdf8b34eeea31bf37dcfad221b4818775fda3b20f"} Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.277266 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" event={"ID":"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6","Type":"ContainerStarted","Data":"ed69eaf9e445234f0e858a52ba50af026ec9ad43f4e70154de73be652d605d02"} Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.277280 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tjsn" event={"ID":"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef","Type":"ContainerStarted","Data":"5847cf43309d5e24d03bc8e4dfc25f7b0751e057a427e42eeee97f646e0336f4"} Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.277295 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m" event={"ID":"962e4f14-b280-4ffa-bde0-6cc4b0d86144","Type":"ContainerDied","Data":"1a1ddd52c91cd0d66eb9ce648f30ab72b5598fdd4aa147bf1505ead58516cba8"} Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.277316 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a1ddd52c91cd0d66eb9ce648f30ab72b5598fdd4aa147bf1505ead58516cba8" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.277446 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.285606 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.276250 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" podStartSLOduration=137.27623637 podStartE2EDuration="2m17.27623637s" podCreationTimestamp="2025-12-01 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:10.252054002 +0000 UTC m=+162.623164413" watchObservedRunningTime="2025-12-01 17:28:10.27623637 +0000 UTC m=+162.647346781" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.298973 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.313659 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.439197 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e2a3315-54d7-4232-af8e-6f4df286fc2f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9e2a3315-54d7-4232-af8e-6f4df286fc2f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.439352 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e2a3315-54d7-4232-af8e-6f4df286fc2f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9e2a3315-54d7-4232-af8e-6f4df286fc2f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.540852 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e2a3315-54d7-4232-af8e-6f4df286fc2f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9e2a3315-54d7-4232-af8e-6f4df286fc2f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.540925 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e2a3315-54d7-4232-af8e-6f4df286fc2f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9e2a3315-54d7-4232-af8e-6f4df286fc2f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.541418 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e2a3315-54d7-4232-af8e-6f4df286fc2f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9e2a3315-54d7-4232-af8e-6f4df286fc2f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.579168 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.582202 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e2a3315-54d7-4232-af8e-6f4df286fc2f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9e2a3315-54d7-4232-af8e-6f4df286fc2f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.596607 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pwdv6"] Dec 01 17:28:10 crc kubenswrapper[4868]: I1201 17:28:10.612742 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 17:28:10 crc kubenswrapper[4868]: W1201 17:28:10.673078 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6332f65b_5f50_4e61_9f2d_b4333fda483b.slice/crio-f2ed393f6a86b1e6a801d8a17df8357df3bd31c9277716bd7778bb07f4a5bdc3 WatchSource:0}: Error finding container f2ed393f6a86b1e6a801d8a17df8357df3bd31c9277716bd7778bb07f4a5bdc3: Status 404 returned error can't find the container with id f2ed393f6a86b1e6a801d8a17df8357df3bd31c9277716bd7778bb07f4a5bdc3 Dec 01 17:28:10 crc kubenswrapper[4868]: W1201 17:28:10.698347 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod77bde072_b3d6_494f_88e4_0732ec5f3ebb.slice/crio-52205a8b34d81b76fd59dee65e233e1995f671f912e3079f32f557025f68ad06 WatchSource:0}: Error finding container 52205a8b34d81b76fd59dee65e233e1995f671f912e3079f32f557025f68ad06: Status 404 returned error can't find the container with id 52205a8b34d81b76fd59dee65e233e1995f671f912e3079f32f557025f68ad06 Dec 01 17:28:11 crc kubenswrapper[4868]: I1201 17:28:11.243875 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 17:28:11 crc kubenswrapper[4868]: I1201 17:28:11.300473 4868 generic.go:334] "Generic (PLEG): container finished" podID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" containerID="2e60ea3c7ad9b897893213af7ecc2484b968c8f0f00e4b4e3be53c02bc483d65" exitCode=0 Dec 01 17:28:11 crc kubenswrapper[4868]: I1201 17:28:11.300574 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tjsn" event={"ID":"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef","Type":"ContainerDied","Data":"2e60ea3c7ad9b897893213af7ecc2484b968c8f0f00e4b4e3be53c02bc483d65"} Dec 01 17:28:11 crc kubenswrapper[4868]: I1201 17:28:11.312010 4868 generic.go:334] "Generic (PLEG): container finished" podID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" containerID="d6d11c99b8802ef5c4d80b716e301094f0a7c3fc00b0ca3a66f411a679c21f07" exitCode=0 Dec 01 17:28:11 crc kubenswrapper[4868]: I1201 17:28:11.312089 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2f8vp" event={"ID":"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e","Type":"ContainerDied","Data":"d6d11c99b8802ef5c4d80b716e301094f0a7c3fc00b0ca3a66f411a679c21f07"} Dec 01 17:28:11 crc kubenswrapper[4868]: I1201 17:28:11.312119 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2f8vp" event={"ID":"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e","Type":"ContainerStarted","Data":"a8a9cc3beaef944fd48ce80d52139b000a3ddab29129ae8fb45f9bfa1de8f6ae"} Dec 01 17:28:11 crc kubenswrapper[4868]: I1201 17:28:11.320456 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"77bde072-b3d6-494f-88e4-0732ec5f3ebb","Type":"ContainerStarted","Data":"52205a8b34d81b76fd59dee65e233e1995f671f912e3079f32f557025f68ad06"} Dec 01 17:28:11 crc kubenswrapper[4868]: I1201 17:28:11.345499 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwdv6" event={"ID":"6332f65b-5f50-4e61-9f2d-b4333fda483b","Type":"ContainerStarted","Data":"cf302c29f1b56ee7be86d0e4fb9989908674e311b4acb83ec0035d1d6b2bf9aa"} Dec 01 17:28:11 crc kubenswrapper[4868]: I1201 17:28:11.349079 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwdv6" event={"ID":"6332f65b-5f50-4e61-9f2d-b4333fda483b","Type":"ContainerStarted","Data":"f2ed393f6a86b1e6a801d8a17df8357df3bd31c9277716bd7778bb07f4a5bdc3"} Dec 01 17:28:12 crc kubenswrapper[4868]: I1201 17:28:12.384293 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9e2a3315-54d7-4232-af8e-6f4df286fc2f","Type":"ContainerStarted","Data":"c291476924eead60bf58ea7566ea19551f94920e646b1de26e18961e0c4f24d6"} Dec 01 17:28:12 crc kubenswrapper[4868]: I1201 17:28:12.393871 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"77bde072-b3d6-494f-88e4-0732ec5f3ebb","Type":"ContainerStarted","Data":"b564ca2478a0708a891d5ebbb4deae81bd718ee8413c33b68589b99757888b46"} Dec 01 17:28:12 crc kubenswrapper[4868]: I1201 17:28:12.415832 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.415811869 podStartE2EDuration="3.415811869s" podCreationTimestamp="2025-12-01 17:28:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:12.413601728 +0000 UTC m=+164.784712139" watchObservedRunningTime="2025-12-01 17:28:12.415811869 +0000 UTC m=+164.786922280" Dec 01 17:28:12 crc kubenswrapper[4868]: I1201 17:28:12.416146 4868 generic.go:334] "Generic (PLEG): container finished" podID="6332f65b-5f50-4e61-9f2d-b4333fda483b" containerID="cf302c29f1b56ee7be86d0e4fb9989908674e311b4acb83ec0035d1d6b2bf9aa" exitCode=0 Dec 01 17:28:12 crc kubenswrapper[4868]: I1201 17:28:12.416219 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwdv6" event={"ID":"6332f65b-5f50-4e61-9f2d-b4333fda483b","Type":"ContainerDied","Data":"cf302c29f1b56ee7be86d0e4fb9989908674e311b4acb83ec0035d1d6b2bf9aa"} Dec 01 17:28:13 crc kubenswrapper[4868]: I1201 17:28:13.492076 4868 generic.go:334] "Generic (PLEG): container finished" podID="77bde072-b3d6-494f-88e4-0732ec5f3ebb" containerID="b564ca2478a0708a891d5ebbb4deae81bd718ee8413c33b68589b99757888b46" exitCode=0 Dec 01 17:28:13 crc kubenswrapper[4868]: I1201 17:28:13.492144 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"77bde072-b3d6-494f-88e4-0732ec5f3ebb","Type":"ContainerDied","Data":"b564ca2478a0708a891d5ebbb4deae81bd718ee8413c33b68589b99757888b46"} Dec 01 17:28:13 crc kubenswrapper[4868]: I1201 17:28:13.495117 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9e2a3315-54d7-4232-af8e-6f4df286fc2f","Type":"ContainerStarted","Data":"ae2904ea0f6d93a99923738e016af4be73c42dd1294ffa1c797ea121bcd3ee15"} Dec 01 17:28:13 crc kubenswrapper[4868]: I1201 17:28:13.538187 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.538158885 podStartE2EDuration="3.538158885s" podCreationTimestamp="2025-12-01 17:28:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:13.530835713 +0000 UTC m=+165.901946124" watchObservedRunningTime="2025-12-01 17:28:13.538158885 +0000 UTC m=+165.909269306" Dec 01 17:28:14 crc kubenswrapper[4868]: I1201 17:28:14.515253 4868 generic.go:334] "Generic (PLEG): container finished" podID="9e2a3315-54d7-4232-af8e-6f4df286fc2f" containerID="ae2904ea0f6d93a99923738e016af4be73c42dd1294ffa1c797ea121bcd3ee15" exitCode=0 Dec 01 17:28:14 crc kubenswrapper[4868]: I1201 17:28:14.516577 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9e2a3315-54d7-4232-af8e-6f4df286fc2f","Type":"ContainerDied","Data":"ae2904ea0f6d93a99923738e016af4be73c42dd1294ffa1c797ea121bcd3ee15"} Dec 01 17:28:15 crc kubenswrapper[4868]: I1201 17:28:15.162570 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 17:28:15 crc kubenswrapper[4868]: I1201 17:28:15.229552 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/77bde072-b3d6-494f-88e4-0732ec5f3ebb-kubelet-dir\") pod \"77bde072-b3d6-494f-88e4-0732ec5f3ebb\" (UID: \"77bde072-b3d6-494f-88e4-0732ec5f3ebb\") " Dec 01 17:28:15 crc kubenswrapper[4868]: I1201 17:28:15.229701 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/77bde072-b3d6-494f-88e4-0732ec5f3ebb-kube-api-access\") pod \"77bde072-b3d6-494f-88e4-0732ec5f3ebb\" (UID: \"77bde072-b3d6-494f-88e4-0732ec5f3ebb\") " Dec 01 17:28:15 crc kubenswrapper[4868]: I1201 17:28:15.230743 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77bde072-b3d6-494f-88e4-0732ec5f3ebb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "77bde072-b3d6-494f-88e4-0732ec5f3ebb" (UID: "77bde072-b3d6-494f-88e4-0732ec5f3ebb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:28:15 crc kubenswrapper[4868]: I1201 17:28:15.246760 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77bde072-b3d6-494f-88e4-0732ec5f3ebb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "77bde072-b3d6-494f-88e4-0732ec5f3ebb" (UID: "77bde072-b3d6-494f-88e4-0732ec5f3ebb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:28:15 crc kubenswrapper[4868]: I1201 17:28:15.332235 4868 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/77bde072-b3d6-494f-88e4-0732ec5f3ebb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 17:28:15 crc kubenswrapper[4868]: I1201 17:28:15.332285 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/77bde072-b3d6-494f-88e4-0732ec5f3ebb-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 17:28:15 crc kubenswrapper[4868]: I1201 17:28:15.605053 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"77bde072-b3d6-494f-88e4-0732ec5f3ebb","Type":"ContainerDied","Data":"52205a8b34d81b76fd59dee65e233e1995f671f912e3079f32f557025f68ad06"} Dec 01 17:28:15 crc kubenswrapper[4868]: I1201 17:28:15.605108 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52205a8b34d81b76fd59dee65e233e1995f671f912e3079f32f557025f68ad06" Dec 01 17:28:15 crc kubenswrapper[4868]: I1201 17:28:15.605184 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 17:28:16 crc kubenswrapper[4868]: I1201 17:28:16.318440 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 17:28:16 crc kubenswrapper[4868]: I1201 17:28:16.354997 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e2a3315-54d7-4232-af8e-6f4df286fc2f-kube-api-access\") pod \"9e2a3315-54d7-4232-af8e-6f4df286fc2f\" (UID: \"9e2a3315-54d7-4232-af8e-6f4df286fc2f\") " Dec 01 17:28:16 crc kubenswrapper[4868]: I1201 17:28:16.355225 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e2a3315-54d7-4232-af8e-6f4df286fc2f-kubelet-dir\") pod \"9e2a3315-54d7-4232-af8e-6f4df286fc2f\" (UID: \"9e2a3315-54d7-4232-af8e-6f4df286fc2f\") " Dec 01 17:28:16 crc kubenswrapper[4868]: I1201 17:28:16.357090 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e2a3315-54d7-4232-af8e-6f4df286fc2f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9e2a3315-54d7-4232-af8e-6f4df286fc2f" (UID: "9e2a3315-54d7-4232-af8e-6f4df286fc2f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:28:16 crc kubenswrapper[4868]: I1201 17:28:16.357659 4868 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e2a3315-54d7-4232-af8e-6f4df286fc2f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 17:28:16 crc kubenswrapper[4868]: I1201 17:28:16.372845 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e2a3315-54d7-4232-af8e-6f4df286fc2f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9e2a3315-54d7-4232-af8e-6f4df286fc2f" (UID: "9e2a3315-54d7-4232-af8e-6f4df286fc2f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:28:16 crc kubenswrapper[4868]: I1201 17:28:16.459123 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e2a3315-54d7-4232-af8e-6f4df286fc2f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 17:28:16 crc kubenswrapper[4868]: I1201 17:28:16.520043 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-nzh9r" Dec 01 17:28:16 crc kubenswrapper[4868]: I1201 17:28:16.663483 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9e2a3315-54d7-4232-af8e-6f4df286fc2f","Type":"ContainerDied","Data":"c291476924eead60bf58ea7566ea19551f94920e646b1de26e18961e0c4f24d6"} Dec 01 17:28:16 crc kubenswrapper[4868]: I1201 17:28:16.663529 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 17:28:16 crc kubenswrapper[4868]: I1201 17:28:16.663541 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c291476924eead60bf58ea7566ea19551f94920e646b1de26e18961e0c4f24d6" Dec 01 17:28:16 crc kubenswrapper[4868]: I1201 17:28:16.773449 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs\") pod \"network-metrics-daemon-nxfkx\" (UID: \"4afd9a0c-e8c7-42ec-9771-036c3441cbce\") " pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:28:16 crc kubenswrapper[4868]: I1201 17:28:16.791888 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4afd9a0c-e8c7-42ec-9771-036c3441cbce-metrics-certs\") pod \"network-metrics-daemon-nxfkx\" (UID: \"4afd9a0c-e8c7-42ec-9771-036c3441cbce\") " pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:28:16 crc kubenswrapper[4868]: I1201 17:28:16.892081 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nxfkx" Dec 01 17:28:17 crc kubenswrapper[4868]: I1201 17:28:17.567056 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-nxfkx"] Dec 01 17:28:17 crc kubenswrapper[4868]: I1201 17:28:17.701017 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" event={"ID":"4afd9a0c-e8c7-42ec-9771-036c3441cbce","Type":"ContainerStarted","Data":"faa403e2c5f1a82aa88dfe16ec5bbc567ad2055b58b937cb977399d6289cbaf6"} Dec 01 17:28:17 crc kubenswrapper[4868]: I1201 17:28:17.885881 4868 patch_prober.go:28] interesting pod/console-f9d7485db-c7pwq container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Dec 01 17:28:17 crc kubenswrapper[4868]: I1201 17:28:17.885962 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-c7pwq" podUID="e90813d0-8960-450f-aaa4-047341c50947" containerName="console" probeResult="failure" output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" Dec 01 17:28:18 crc kubenswrapper[4868]: I1201 17:28:18.077714 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:18 crc kubenswrapper[4868]: I1201 17:28:18.077743 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:18 crc kubenswrapper[4868]: I1201 17:28:18.077793 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:18 crc kubenswrapper[4868]: I1201 17:28:18.077818 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:18 crc kubenswrapper[4868]: I1201 17:28:18.730333 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" event={"ID":"4afd9a0c-e8c7-42ec-9771-036c3441cbce","Type":"ContainerStarted","Data":"7feb07ef4b55d1ce0a649bc7506d5619e542e4da8a2bed4b43bea3b329b62061"} Dec 01 17:28:19 crc kubenswrapper[4868]: I1201 17:28:19.759121 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nxfkx" event={"ID":"4afd9a0c-e8c7-42ec-9771-036c3441cbce","Type":"ContainerStarted","Data":"7e77b842220bb4116daf71b6e16644461b10e4803a891057614cf0cc5585b29b"} Dec 01 17:28:19 crc kubenswrapper[4868]: I1201 17:28:19.787724 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-nxfkx" podStartSLOduration=145.787697958 podStartE2EDuration="2m25.787697958s" podCreationTimestamp="2025-12-01 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:19.780697745 +0000 UTC m=+172.151808156" watchObservedRunningTime="2025-12-01 17:28:19.787697958 +0000 UTC m=+172.158808369" Dec 01 17:28:25 crc kubenswrapper[4868]: I1201 17:28:25.905027 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:28:25 crc kubenswrapper[4868]: I1201 17:28:25.906032 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:28:27 crc kubenswrapper[4868]: I1201 17:28:27.891502 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:28:27 crc kubenswrapper[4868]: I1201 17:28:27.897325 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:28:28 crc kubenswrapper[4868]: I1201 17:28:28.032082 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:28:28 crc kubenswrapper[4868]: I1201 17:28:28.121953 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:28 crc kubenswrapper[4868]: I1201 17:28:28.122880 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:28 crc kubenswrapper[4868]: I1201 17:28:28.121961 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:28 crc kubenswrapper[4868]: I1201 17:28:28.123596 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:28 crc kubenswrapper[4868]: I1201 17:28:28.123696 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-qnznl" Dec 01 17:28:28 crc kubenswrapper[4868]: I1201 17:28:28.124241 4868 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"8be494799fa1ca47fef75ee37d25f33c1f9061e74adb238d211d708d8e866faf"} pod="openshift-console/downloads-7954f5f757-qnznl" containerMessage="Container download-server failed liveness probe, will be restarted" Dec 01 17:28:28 crc kubenswrapper[4868]: I1201 17:28:28.124413 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" containerID="cri-o://8be494799fa1ca47fef75ee37d25f33c1f9061e74adb238d211d708d8e866faf" gracePeriod=2 Dec 01 17:28:28 crc kubenswrapper[4868]: I1201 17:28:28.125006 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:28 crc kubenswrapper[4868]: I1201 17:28:28.125104 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:28 crc kubenswrapper[4868]: I1201 17:28:28.893774 4868 generic.go:334] "Generic (PLEG): container finished" podID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerID="8be494799fa1ca47fef75ee37d25f33c1f9061e74adb238d211d708d8e866faf" exitCode=0 Dec 01 17:28:28 crc kubenswrapper[4868]: I1201 17:28:28.893887 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qnznl" event={"ID":"9307075d-5039-4f41-ac9a-8d6c2dbca531","Type":"ContainerDied","Data":"8be494799fa1ca47fef75ee37d25f33c1f9061e74adb238d211d708d8e866faf"} Dec 01 17:28:36 crc kubenswrapper[4868]: I1201 17:28:36.832295 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 17:28:38 crc kubenswrapper[4868]: I1201 17:28:38.078199 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:38 crc kubenswrapper[4868]: I1201 17:28:38.078277 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:38 crc kubenswrapper[4868]: I1201 17:28:38.107071 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4njnk" Dec 01 17:28:44 crc kubenswrapper[4868]: I1201 17:28:44.807048 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 17:28:44 crc kubenswrapper[4868]: E1201 17:28:44.807998 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77bde072-b3d6-494f-88e4-0732ec5f3ebb" containerName="pruner" Dec 01 17:28:44 crc kubenswrapper[4868]: I1201 17:28:44.808012 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="77bde072-b3d6-494f-88e4-0732ec5f3ebb" containerName="pruner" Dec 01 17:28:44 crc kubenswrapper[4868]: E1201 17:28:44.808027 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e2a3315-54d7-4232-af8e-6f4df286fc2f" containerName="pruner" Dec 01 17:28:44 crc kubenswrapper[4868]: I1201 17:28:44.808034 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e2a3315-54d7-4232-af8e-6f4df286fc2f" containerName="pruner" Dec 01 17:28:44 crc kubenswrapper[4868]: I1201 17:28:44.808159 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e2a3315-54d7-4232-af8e-6f4df286fc2f" containerName="pruner" Dec 01 17:28:44 crc kubenswrapper[4868]: I1201 17:28:44.808175 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="77bde072-b3d6-494f-88e4-0732ec5f3ebb" containerName="pruner" Dec 01 17:28:44 crc kubenswrapper[4868]: I1201 17:28:44.808597 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 17:28:44 crc kubenswrapper[4868]: I1201 17:28:44.810810 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 17:28:44 crc kubenswrapper[4868]: I1201 17:28:44.811820 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 17:28:44 crc kubenswrapper[4868]: I1201 17:28:44.814086 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 17:28:44 crc kubenswrapper[4868]: I1201 17:28:44.909639 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da15d531-bc43-41f0-a568-a6a34308cfa3-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"da15d531-bc43-41f0-a568-a6a34308cfa3\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 17:28:44 crc kubenswrapper[4868]: I1201 17:28:44.909696 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/da15d531-bc43-41f0-a568-a6a34308cfa3-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"da15d531-bc43-41f0-a568-a6a34308cfa3\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 17:28:45 crc kubenswrapper[4868]: I1201 17:28:45.011388 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/da15d531-bc43-41f0-a568-a6a34308cfa3-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"da15d531-bc43-41f0-a568-a6a34308cfa3\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 17:28:45 crc kubenswrapper[4868]: I1201 17:28:45.011561 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da15d531-bc43-41f0-a568-a6a34308cfa3-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"da15d531-bc43-41f0-a568-a6a34308cfa3\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 17:28:45 crc kubenswrapper[4868]: I1201 17:28:45.011599 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/da15d531-bc43-41f0-a568-a6a34308cfa3-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"da15d531-bc43-41f0-a568-a6a34308cfa3\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 17:28:45 crc kubenswrapper[4868]: I1201 17:28:45.039082 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da15d531-bc43-41f0-a568-a6a34308cfa3-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"da15d531-bc43-41f0-a568-a6a34308cfa3\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 17:28:45 crc kubenswrapper[4868]: I1201 17:28:45.142067 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 17:28:45 crc kubenswrapper[4868]: E1201 17:28:45.690639 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 01 17:28:45 crc kubenswrapper[4868]: E1201 17:28:45.691922 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pvxwm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xtcsm_openshift-marketplace(b1529d6b-804b-4fb6-89a2-53cb5e2c5f06): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 17:28:45 crc kubenswrapper[4868]: E1201 17:28:45.693450 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-xtcsm" podUID="b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" Dec 01 17:28:48 crc kubenswrapper[4868]: I1201 17:28:48.078853 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:48 crc kubenswrapper[4868]: I1201 17:28:48.078987 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:48 crc kubenswrapper[4868]: E1201 17:28:48.271305 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xtcsm" podUID="b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" Dec 01 17:28:48 crc kubenswrapper[4868]: E1201 17:28:48.378610 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 01 17:28:48 crc kubenswrapper[4868]: E1201 17:28:48.378951 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n7qrp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-b89j6_openshift-marketplace(262076bc-ef10-4531-a6f9-f32148ee7a3e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 17:28:48 crc kubenswrapper[4868]: E1201 17:28:48.380022 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-b89j6" podUID="262076bc-ef10-4531-a6f9-f32148ee7a3e" Dec 01 17:28:50 crc kubenswrapper[4868]: E1201 17:28:50.012425 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-b89j6" podUID="262076bc-ef10-4531-a6f9-f32148ee7a3e" Dec 01 17:28:50 crc kubenswrapper[4868]: E1201 17:28:50.078026 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 01 17:28:50 crc kubenswrapper[4868]: E1201 17:28:50.078273 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zlbmh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-4tjsn_openshift-marketplace(14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 17:28:50 crc kubenswrapper[4868]: E1201 17:28:50.079540 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-4tjsn" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" Dec 01 17:28:50 crc kubenswrapper[4868]: I1201 17:28:50.610134 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 17:28:50 crc kubenswrapper[4868]: I1201 17:28:50.611885 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 17:28:50 crc kubenswrapper[4868]: I1201 17:28:50.614512 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 17:28:50 crc kubenswrapper[4868]: I1201 17:28:50.799919 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed9731d5-c84c-4e48-91c4-2e66fba896a6-kube-api-access\") pod \"installer-9-crc\" (UID: \"ed9731d5-c84c-4e48-91c4-2e66fba896a6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 17:28:50 crc kubenswrapper[4868]: I1201 17:28:50.800040 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed9731d5-c84c-4e48-91c4-2e66fba896a6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ed9731d5-c84c-4e48-91c4-2e66fba896a6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 17:28:50 crc kubenswrapper[4868]: I1201 17:28:50.800075 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ed9731d5-c84c-4e48-91c4-2e66fba896a6-var-lock\") pod \"installer-9-crc\" (UID: \"ed9731d5-c84c-4e48-91c4-2e66fba896a6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 17:28:50 crc kubenswrapper[4868]: I1201 17:28:50.901283 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed9731d5-c84c-4e48-91c4-2e66fba896a6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ed9731d5-c84c-4e48-91c4-2e66fba896a6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 17:28:50 crc kubenswrapper[4868]: I1201 17:28:50.901364 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ed9731d5-c84c-4e48-91c4-2e66fba896a6-var-lock\") pod \"installer-9-crc\" (UID: \"ed9731d5-c84c-4e48-91c4-2e66fba896a6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 17:28:50 crc kubenswrapper[4868]: I1201 17:28:50.901422 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed9731d5-c84c-4e48-91c4-2e66fba896a6-kube-api-access\") pod \"installer-9-crc\" (UID: \"ed9731d5-c84c-4e48-91c4-2e66fba896a6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 17:28:50 crc kubenswrapper[4868]: I1201 17:28:50.901459 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed9731d5-c84c-4e48-91c4-2e66fba896a6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ed9731d5-c84c-4e48-91c4-2e66fba896a6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 17:28:50 crc kubenswrapper[4868]: I1201 17:28:50.901637 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ed9731d5-c84c-4e48-91c4-2e66fba896a6-var-lock\") pod \"installer-9-crc\" (UID: \"ed9731d5-c84c-4e48-91c4-2e66fba896a6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 17:28:50 crc kubenswrapper[4868]: I1201 17:28:50.927424 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed9731d5-c84c-4e48-91c4-2e66fba896a6-kube-api-access\") pod \"installer-9-crc\" (UID: \"ed9731d5-c84c-4e48-91c4-2e66fba896a6\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 17:28:50 crc kubenswrapper[4868]: I1201 17:28:50.940894 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.084561 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-4tjsn" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.186784 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.187337 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jwg6w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-2f8vp_openshift-marketplace(1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.188542 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-2f8vp" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.201790 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.201991 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c64j5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-fnrd7_openshift-marketplace(de9baf7b-7da5-4cd3-908a-d3bc26f918f1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.203282 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-fnrd7" podUID="de9baf7b-7da5-4cd3-908a-d3bc26f918f1" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.217414 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.218295 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pgls8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-pwdv6_openshift-marketplace(6332f65b-5f50-4e61-9f2d-b4333fda483b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.219158 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.219709 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-pwdv6" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.220560 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4b8mw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-c9tpp_openshift-marketplace(b2c7b9b4-a824-4639-9dfa-21778b900c9a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.222224 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-c9tpp" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.249182 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.249409 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n2jhx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-hz4pc_openshift-marketplace(d7c0aa99-172b-41b7-80e5-2cbc6532cae1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 17:28:54 crc kubenswrapper[4868]: E1201 17:28:54.250828 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-hz4pc" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" Dec 01 17:28:54 crc kubenswrapper[4868]: I1201 17:28:54.413113 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 17:28:54 crc kubenswrapper[4868]: W1201 17:28:54.419897 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podda15d531_bc43_41f0_a568_a6a34308cfa3.slice/crio-0a8d06f1272e2364fc6743097faab368363cc2360c8b273e1fb2f276040913db WatchSource:0}: Error finding container 0a8d06f1272e2364fc6743097faab368363cc2360c8b273e1fb2f276040913db: Status 404 returned error can't find the container with id 0a8d06f1272e2364fc6743097faab368363cc2360c8b273e1fb2f276040913db Dec 01 17:28:54 crc kubenswrapper[4868]: I1201 17:28:54.576586 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 17:28:55 crc kubenswrapper[4868]: I1201 17:28:55.075725 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qnznl" event={"ID":"9307075d-5039-4f41-ac9a-8d6c2dbca531","Type":"ContainerStarted","Data":"938c90430b1cb9ab879bca0da23dba82a33225fec41cf3dcf1b7d7a76ae539df"} Dec 01 17:28:55 crc kubenswrapper[4868]: I1201 17:28:55.076224 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-qnznl" Dec 01 17:28:55 crc kubenswrapper[4868]: I1201 17:28:55.078072 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:55 crc kubenswrapper[4868]: I1201 17:28:55.078149 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:55 crc kubenswrapper[4868]: I1201 17:28:55.079844 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ed9731d5-c84c-4e48-91c4-2e66fba896a6","Type":"ContainerStarted","Data":"4acf6c8a2136f0b65fce38151339a90aa6f197fc5aae62ff6aba2dae87234166"} Dec 01 17:28:55 crc kubenswrapper[4868]: I1201 17:28:55.080073 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ed9731d5-c84c-4e48-91c4-2e66fba896a6","Type":"ContainerStarted","Data":"41f9965470ced10dd5b00144f999cee53bfeff593ae6a9f00955b7106016aef7"} Dec 01 17:28:55 crc kubenswrapper[4868]: I1201 17:28:55.082391 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"da15d531-bc43-41f0-a568-a6a34308cfa3","Type":"ContainerStarted","Data":"c4fda5a8329b0a69e7535595524aa19bb57da72fb592ae1a43ad23d5acff19bd"} Dec 01 17:28:55 crc kubenswrapper[4868]: I1201 17:28:55.082433 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"da15d531-bc43-41f0-a568-a6a34308cfa3","Type":"ContainerStarted","Data":"0a8d06f1272e2364fc6743097faab368363cc2360c8b273e1fb2f276040913db"} Dec 01 17:28:55 crc kubenswrapper[4868]: E1201 17:28:55.083913 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-c9tpp" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" Dec 01 17:28:55 crc kubenswrapper[4868]: E1201 17:28:55.085075 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-fnrd7" podUID="de9baf7b-7da5-4cd3-908a-d3bc26f918f1" Dec 01 17:28:55 crc kubenswrapper[4868]: E1201 17:28:55.085613 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-hz4pc" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" Dec 01 17:28:55 crc kubenswrapper[4868]: E1201 17:28:55.085691 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-2f8vp" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" Dec 01 17:28:55 crc kubenswrapper[4868]: E1201 17:28:55.087189 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-pwdv6" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" Dec 01 17:28:55 crc kubenswrapper[4868]: I1201 17:28:55.180481 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=5.180454138 podStartE2EDuration="5.180454138s" podCreationTimestamp="2025-12-01 17:28:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:55.150690026 +0000 UTC m=+207.521800437" watchObservedRunningTime="2025-12-01 17:28:55.180454138 +0000 UTC m=+207.551564549" Dec 01 17:28:55 crc kubenswrapper[4868]: I1201 17:28:55.220933 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=11.220907273 podStartE2EDuration="11.220907273s" podCreationTimestamp="2025-12-01 17:28:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:28:55.197073218 +0000 UTC m=+207.568183629" watchObservedRunningTime="2025-12-01 17:28:55.220907273 +0000 UTC m=+207.592017694" Dec 01 17:28:55 crc kubenswrapper[4868]: I1201 17:28:55.907117 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:28:55 crc kubenswrapper[4868]: I1201 17:28:55.907172 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:28:55 crc kubenswrapper[4868]: I1201 17:28:55.907224 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:28:55 crc kubenswrapper[4868]: I1201 17:28:55.907894 4868 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd"} pod="openshift-machine-config-operator/machine-config-daemon-njgpn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 17:28:55 crc kubenswrapper[4868]: I1201 17:28:55.907974 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" containerID="cri-o://fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd" gracePeriod=600 Dec 01 17:28:56 crc kubenswrapper[4868]: I1201 17:28:56.089103 4868 generic.go:334] "Generic (PLEG): container finished" podID="da15d531-bc43-41f0-a568-a6a34308cfa3" containerID="c4fda5a8329b0a69e7535595524aa19bb57da72fb592ae1a43ad23d5acff19bd" exitCode=0 Dec 01 17:28:56 crc kubenswrapper[4868]: I1201 17:28:56.089166 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"da15d531-bc43-41f0-a568-a6a34308cfa3","Type":"ContainerDied","Data":"c4fda5a8329b0a69e7535595524aa19bb57da72fb592ae1a43ad23d5acff19bd"} Dec 01 17:28:56 crc kubenswrapper[4868]: I1201 17:28:56.090877 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:56 crc kubenswrapper[4868]: I1201 17:28:56.090932 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:57 crc kubenswrapper[4868]: I1201 17:28:57.098714 4868 generic.go:334] "Generic (PLEG): container finished" podID="ec4e08bf-df56-4490-8e79-a4175c727405" containerID="fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd" exitCode=0 Dec 01 17:28:57 crc kubenswrapper[4868]: I1201 17:28:57.099752 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerDied","Data":"fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd"} Dec 01 17:28:57 crc kubenswrapper[4868]: I1201 17:28:57.099790 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"5a50633c9f5f1a502388b67f58e02c5f45fd0e84eeadf70b339824d2940a6594"} Dec 01 17:28:57 crc kubenswrapper[4868]: I1201 17:28:57.324452 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 17:28:57 crc kubenswrapper[4868]: I1201 17:28:57.510257 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/da15d531-bc43-41f0-a568-a6a34308cfa3-kubelet-dir\") pod \"da15d531-bc43-41f0-a568-a6a34308cfa3\" (UID: \"da15d531-bc43-41f0-a568-a6a34308cfa3\") " Dec 01 17:28:57 crc kubenswrapper[4868]: I1201 17:28:57.510419 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da15d531-bc43-41f0-a568-a6a34308cfa3-kube-api-access\") pod \"da15d531-bc43-41f0-a568-a6a34308cfa3\" (UID: \"da15d531-bc43-41f0-a568-a6a34308cfa3\") " Dec 01 17:28:57 crc kubenswrapper[4868]: I1201 17:28:57.511879 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da15d531-bc43-41f0-a568-a6a34308cfa3-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "da15d531-bc43-41f0-a568-a6a34308cfa3" (UID: "da15d531-bc43-41f0-a568-a6a34308cfa3"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:28:57 crc kubenswrapper[4868]: I1201 17:28:57.524790 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da15d531-bc43-41f0-a568-a6a34308cfa3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "da15d531-bc43-41f0-a568-a6a34308cfa3" (UID: "da15d531-bc43-41f0-a568-a6a34308cfa3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:28:57 crc kubenswrapper[4868]: I1201 17:28:57.612215 4868 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/da15d531-bc43-41f0-a568-a6a34308cfa3-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 17:28:57 crc kubenswrapper[4868]: I1201 17:28:57.612264 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da15d531-bc43-41f0-a568-a6a34308cfa3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 17:28:58 crc kubenswrapper[4868]: I1201 17:28:58.078384 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:58 crc kubenswrapper[4868]: I1201 17:28:58.078398 4868 patch_prober.go:28] interesting pod/downloads-7954f5f757-qnznl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Dec 01 17:28:58 crc kubenswrapper[4868]: I1201 17:28:58.078470 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:58 crc kubenswrapper[4868]: I1201 17:28:58.078518 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-qnznl" podUID="9307075d-5039-4f41-ac9a-8d6c2dbca531" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.23:8080/\": dial tcp 10.217.0.23:8080: connect: connection refused" Dec 01 17:28:58 crc kubenswrapper[4868]: I1201 17:28:58.105187 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"da15d531-bc43-41f0-a568-a6a34308cfa3","Type":"ContainerDied","Data":"0a8d06f1272e2364fc6743097faab368363cc2360c8b273e1fb2f276040913db"} Dec 01 17:28:58 crc kubenswrapper[4868]: I1201 17:28:58.105234 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a8d06f1272e2364fc6743097faab368363cc2360c8b273e1fb2f276040913db" Dec 01 17:28:58 crc kubenswrapper[4868]: I1201 17:28:58.105261 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 17:29:04 crc kubenswrapper[4868]: I1201 17:29:04.143501 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xtcsm" event={"ID":"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06","Type":"ContainerStarted","Data":"2e84e511a4b3d6e3bc4fde7cc484e08f0239dad6a9374c13c8dd6e73a2bdb1c8"} Dec 01 17:29:04 crc kubenswrapper[4868]: I1201 17:29:04.146431 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b89j6" event={"ID":"262076bc-ef10-4531-a6f9-f32148ee7a3e","Type":"ContainerStarted","Data":"e63e969ae8ff94477b7a6ceb39d0e7e13188df53cafdc2f44ebea9c8f4a0e176"} Dec 01 17:29:05 crc kubenswrapper[4868]: I1201 17:29:05.154763 4868 generic.go:334] "Generic (PLEG): container finished" podID="262076bc-ef10-4531-a6f9-f32148ee7a3e" containerID="e63e969ae8ff94477b7a6ceb39d0e7e13188df53cafdc2f44ebea9c8f4a0e176" exitCode=0 Dec 01 17:29:05 crc kubenswrapper[4868]: I1201 17:29:05.154852 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b89j6" event={"ID":"262076bc-ef10-4531-a6f9-f32148ee7a3e","Type":"ContainerDied","Data":"e63e969ae8ff94477b7a6ceb39d0e7e13188df53cafdc2f44ebea9c8f4a0e176"} Dec 01 17:29:05 crc kubenswrapper[4868]: I1201 17:29:05.160550 4868 generic.go:334] "Generic (PLEG): container finished" podID="b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" containerID="2e84e511a4b3d6e3bc4fde7cc484e08f0239dad6a9374c13c8dd6e73a2bdb1c8" exitCode=0 Dec 01 17:29:05 crc kubenswrapper[4868]: I1201 17:29:05.160617 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xtcsm" event={"ID":"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06","Type":"ContainerDied","Data":"2e84e511a4b3d6e3bc4fde7cc484e08f0239dad6a9374c13c8dd6e73a2bdb1c8"} Dec 01 17:29:08 crc kubenswrapper[4868]: I1201 17:29:08.087852 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-qnznl" Dec 01 17:29:19 crc kubenswrapper[4868]: I1201 17:29:19.261712 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2f8vp" event={"ID":"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e","Type":"ContainerStarted","Data":"b46685901acab86a775f40a7851a7fa0412e42b9eb1d66c2b21931cb4b6ab4b5"} Dec 01 17:29:19 crc kubenswrapper[4868]: I1201 17:29:19.282176 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b89j6" event={"ID":"262076bc-ef10-4531-a6f9-f32148ee7a3e","Type":"ContainerStarted","Data":"00136348a5d6fe9d51d6e34c1e88a69b9216fb2ac25f3f9950a8b170dee2edb7"} Dec 01 17:29:19 crc kubenswrapper[4868]: I1201 17:29:19.289393 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xtcsm" event={"ID":"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06","Type":"ContainerStarted","Data":"c2d7039d55006e07b57b82c23ad7262ead3c43e41c20b54382e22e41b5f8c8da"} Dec 01 17:29:19 crc kubenswrapper[4868]: I1201 17:29:19.293203 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwdv6" event={"ID":"6332f65b-5f50-4e61-9f2d-b4333fda483b","Type":"ContainerStarted","Data":"747f002d75adf979a20407ec8852e9555659ca1af2c7068335be9a21848d29ca"} Dec 01 17:29:19 crc kubenswrapper[4868]: I1201 17:29:19.296242 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9tpp" event={"ID":"b2c7b9b4-a824-4639-9dfa-21778b900c9a","Type":"ContainerStarted","Data":"e8919dbc7948177a59fe00b4211e5c84e9c7b11b4c8fa41f0868b28933392fe4"} Dec 01 17:29:19 crc kubenswrapper[4868]: I1201 17:29:19.299528 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnrd7" event={"ID":"de9baf7b-7da5-4cd3-908a-d3bc26f918f1","Type":"ContainerStarted","Data":"3207bdbc0f84b42cdc537b8bffcc1001bcca64e782d5115aecbfca764eb604a3"} Dec 01 17:29:19 crc kubenswrapper[4868]: I1201 17:29:19.301040 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hz4pc" event={"ID":"d7c0aa99-172b-41b7-80e5-2cbc6532cae1","Type":"ContainerStarted","Data":"adf50d549d08c57b185d0a1fb9bd1dcbeb37dcdc1e20be21468322c3fa1e4f42"} Dec 01 17:29:19 crc kubenswrapper[4868]: I1201 17:29:19.302332 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tjsn" event={"ID":"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef","Type":"ContainerStarted","Data":"bc11c670e2c0d188d4cbe9eb7386bd5b695a99f3f44f959f0b45af2813f03135"} Dec 01 17:29:19 crc kubenswrapper[4868]: I1201 17:29:19.326333 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xtcsm" podStartSLOduration=8.139595134 podStartE2EDuration="1m14.326314564s" podCreationTimestamp="2025-12-01 17:28:05 +0000 UTC" firstStartedPulling="2025-12-01 17:28:07.979538091 +0000 UTC m=+160.350648502" lastFinishedPulling="2025-12-01 17:29:14.166257481 +0000 UTC m=+226.537367932" observedRunningTime="2025-12-01 17:29:19.321511739 +0000 UTC m=+231.692622160" watchObservedRunningTime="2025-12-01 17:29:19.326314564 +0000 UTC m=+231.697424975" Dec 01 17:29:19 crc kubenswrapper[4868]: I1201 17:29:19.366073 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b89j6" podStartSLOduration=3.863492279 podStartE2EDuration="1m14.366048879s" podCreationTimestamp="2025-12-01 17:28:05 +0000 UTC" firstStartedPulling="2025-12-01 17:28:07.959004044 +0000 UTC m=+160.330114455" lastFinishedPulling="2025-12-01 17:29:18.461560644 +0000 UTC m=+230.832671055" observedRunningTime="2025-12-01 17:29:19.362747725 +0000 UTC m=+231.733858146" watchObservedRunningTime="2025-12-01 17:29:19.366048879 +0000 UTC m=+231.737159290" Dec 01 17:29:20 crc kubenswrapper[4868]: I1201 17:29:20.547742 4868 generic.go:334] "Generic (PLEG): container finished" podID="de9baf7b-7da5-4cd3-908a-d3bc26f918f1" containerID="3207bdbc0f84b42cdc537b8bffcc1001bcca64e782d5115aecbfca764eb604a3" exitCode=0 Dec 01 17:29:20 crc kubenswrapper[4868]: I1201 17:29:20.547822 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnrd7" event={"ID":"de9baf7b-7da5-4cd3-908a-d3bc26f918f1","Type":"ContainerDied","Data":"3207bdbc0f84b42cdc537b8bffcc1001bcca64e782d5115aecbfca764eb604a3"} Dec 01 17:29:20 crc kubenswrapper[4868]: I1201 17:29:20.614037 4868 generic.go:334] "Generic (PLEG): container finished" podID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" containerID="bc11c670e2c0d188d4cbe9eb7386bd5b695a99f3f44f959f0b45af2813f03135" exitCode=0 Dec 01 17:29:20 crc kubenswrapper[4868]: I1201 17:29:20.614088 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tjsn" event={"ID":"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef","Type":"ContainerDied","Data":"bc11c670e2c0d188d4cbe9eb7386bd5b695a99f3f44f959f0b45af2813f03135"} Dec 01 17:29:21 crc kubenswrapper[4868]: I1201 17:29:21.642474 4868 generic.go:334] "Generic (PLEG): container finished" podID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" containerID="e8919dbc7948177a59fe00b4211e5c84e9c7b11b4c8fa41f0868b28933392fe4" exitCode=0 Dec 01 17:29:21 crc kubenswrapper[4868]: I1201 17:29:21.642552 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9tpp" event={"ID":"b2c7b9b4-a824-4639-9dfa-21778b900c9a","Type":"ContainerDied","Data":"e8919dbc7948177a59fe00b4211e5c84e9c7b11b4c8fa41f0868b28933392fe4"} Dec 01 17:29:21 crc kubenswrapper[4868]: I1201 17:29:21.646777 4868 generic.go:334] "Generic (PLEG): container finished" podID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" containerID="adf50d549d08c57b185d0a1fb9bd1dcbeb37dcdc1e20be21468322c3fa1e4f42" exitCode=0 Dec 01 17:29:21 crc kubenswrapper[4868]: I1201 17:29:21.646848 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hz4pc" event={"ID":"d7c0aa99-172b-41b7-80e5-2cbc6532cae1","Type":"ContainerDied","Data":"adf50d549d08c57b185d0a1fb9bd1dcbeb37dcdc1e20be21468322c3fa1e4f42"} Dec 01 17:29:22 crc kubenswrapper[4868]: I1201 17:29:22.656511 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tjsn" event={"ID":"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef","Type":"ContainerStarted","Data":"f085b3894b9b09f6dcc72340cb60a13cbc958ba4634f06de12a407cbb6d89274"} Dec 01 17:29:22 crc kubenswrapper[4868]: I1201 17:29:22.659495 4868 generic.go:334] "Generic (PLEG): container finished" podID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" containerID="b46685901acab86a775f40a7851a7fa0412e42b9eb1d66c2b21931cb4b6ab4b5" exitCode=0 Dec 01 17:29:22 crc kubenswrapper[4868]: I1201 17:29:22.659549 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2f8vp" event={"ID":"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e","Type":"ContainerDied","Data":"b46685901acab86a775f40a7851a7fa0412e42b9eb1d66c2b21931cb4b6ab4b5"} Dec 01 17:29:22 crc kubenswrapper[4868]: I1201 17:29:22.662159 4868 generic.go:334] "Generic (PLEG): container finished" podID="6332f65b-5f50-4e61-9f2d-b4333fda483b" containerID="747f002d75adf979a20407ec8852e9555659ca1af2c7068335be9a21848d29ca" exitCode=0 Dec 01 17:29:22 crc kubenswrapper[4868]: I1201 17:29:22.662261 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwdv6" event={"ID":"6332f65b-5f50-4e61-9f2d-b4333fda483b","Type":"ContainerDied","Data":"747f002d75adf979a20407ec8852e9555659ca1af2c7068335be9a21848d29ca"} Dec 01 17:29:22 crc kubenswrapper[4868]: I1201 17:29:22.666611 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9tpp" event={"ID":"b2c7b9b4-a824-4639-9dfa-21778b900c9a","Type":"ContainerStarted","Data":"6c87d18bcef7c3d59b130c2225c87b56d162d5ac9e5ce3583898a2080579b5b5"} Dec 01 17:29:22 crc kubenswrapper[4868]: I1201 17:29:22.669319 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnrd7" event={"ID":"de9baf7b-7da5-4cd3-908a-d3bc26f918f1","Type":"ContainerStarted","Data":"cc55d48ae37cb8986ddf62f4b7e17efc0b435d1be46faefeb4f8c57dcc5f5352"} Dec 01 17:29:22 crc kubenswrapper[4868]: I1201 17:29:22.673398 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hz4pc" event={"ID":"d7c0aa99-172b-41b7-80e5-2cbc6532cae1","Type":"ContainerStarted","Data":"2cea147eb14fd0d23bcf87da73c105e7b8c83cea691afa775d90d6cc085c7348"} Dec 01 17:29:22 crc kubenswrapper[4868]: I1201 17:29:22.692834 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4tjsn" podStartSLOduration=4.681338951 podStartE2EDuration="1m14.692816643s" podCreationTimestamp="2025-12-01 17:28:08 +0000 UTC" firstStartedPulling="2025-12-01 17:28:11.308627761 +0000 UTC m=+163.679738172" lastFinishedPulling="2025-12-01 17:29:21.320105453 +0000 UTC m=+233.691215864" observedRunningTime="2025-12-01 17:29:22.689109159 +0000 UTC m=+235.060219580" watchObservedRunningTime="2025-12-01 17:29:22.692816643 +0000 UTC m=+235.063927054" Dec 01 17:29:22 crc kubenswrapper[4868]: I1201 17:29:22.734410 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fnrd7" podStartSLOduration=4.684227739 podStartE2EDuration="1m15.734388519s" podCreationTimestamp="2025-12-01 17:28:07 +0000 UTC" firstStartedPulling="2025-12-01 17:28:10.198483462 +0000 UTC m=+162.569593873" lastFinishedPulling="2025-12-01 17:29:21.248644242 +0000 UTC m=+233.619754653" observedRunningTime="2025-12-01 17:29:22.73158244 +0000 UTC m=+235.102692851" watchObservedRunningTime="2025-12-01 17:29:22.734388519 +0000 UTC m=+235.105498920" Dec 01 17:29:22 crc kubenswrapper[4868]: I1201 17:29:22.784554 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c9tpp" podStartSLOduration=3.782846114 podStartE2EDuration="1m16.784529738s" podCreationTimestamp="2025-12-01 17:28:06 +0000 UTC" firstStartedPulling="2025-12-01 17:28:09.141219864 +0000 UTC m=+161.512330275" lastFinishedPulling="2025-12-01 17:29:22.142903498 +0000 UTC m=+234.514013899" observedRunningTime="2025-12-01 17:29:22.782666135 +0000 UTC m=+235.153776546" watchObservedRunningTime="2025-12-01 17:29:22.784529738 +0000 UTC m=+235.155640149" Dec 01 17:29:22 crc kubenswrapper[4868]: I1201 17:29:22.884003 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hz4pc" podStartSLOduration=3.843279833 podStartE2EDuration="1m16.88397065s" podCreationTimestamp="2025-12-01 17:28:06 +0000 UTC" firstStartedPulling="2025-12-01 17:28:09.141222694 +0000 UTC m=+161.512333105" lastFinishedPulling="2025-12-01 17:29:22.181913501 +0000 UTC m=+234.553023922" observedRunningTime="2025-12-01 17:29:22.881096649 +0000 UTC m=+235.252207060" watchObservedRunningTime="2025-12-01 17:29:22.88397065 +0000 UTC m=+235.255081071" Dec 01 17:29:23 crc kubenswrapper[4868]: I1201 17:29:23.690287 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwdv6" event={"ID":"6332f65b-5f50-4e61-9f2d-b4333fda483b","Type":"ContainerStarted","Data":"b3b78e31c529d875946a4c9c212fdbbff5fdd7825c32083e081be87ff543b7bd"} Dec 01 17:29:23 crc kubenswrapper[4868]: I1201 17:29:23.711005 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pwdv6" podStartSLOduration=3.971701727 podStartE2EDuration="1m14.710979665s" podCreationTimestamp="2025-12-01 17:28:09 +0000 UTC" firstStartedPulling="2025-12-01 17:28:12.426796742 +0000 UTC m=+164.797907153" lastFinishedPulling="2025-12-01 17:29:23.16607468 +0000 UTC m=+235.537185091" observedRunningTime="2025-12-01 17:29:23.708582957 +0000 UTC m=+236.079693378" watchObservedRunningTime="2025-12-01 17:29:23.710979665 +0000 UTC m=+236.082090076" Dec 01 17:29:24 crc kubenswrapper[4868]: I1201 17:29:24.699578 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2f8vp" event={"ID":"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e","Type":"ContainerStarted","Data":"860fb316d47a18c7681dea9bb1438f501a3657b1fc3c566a1a43b02eb94837af"} Dec 01 17:29:24 crc kubenswrapper[4868]: I1201 17:29:24.723820 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2f8vp" podStartSLOduration=3.565247791 podStartE2EDuration="1m15.723795624s" podCreationTimestamp="2025-12-01 17:28:09 +0000 UTC" firstStartedPulling="2025-12-01 17:28:11.377671458 +0000 UTC m=+163.748781869" lastFinishedPulling="2025-12-01 17:29:23.536219291 +0000 UTC m=+235.907329702" observedRunningTime="2025-12-01 17:29:24.722561299 +0000 UTC m=+237.093671710" watchObservedRunningTime="2025-12-01 17:29:24.723795624 +0000 UTC m=+237.094906035" Dec 01 17:29:26 crc kubenswrapper[4868]: I1201 17:29:26.211180 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:29:26 crc kubenswrapper[4868]: I1201 17:29:26.211711 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:29:26 crc kubenswrapper[4868]: I1201 17:29:26.613032 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:29:26 crc kubenswrapper[4868]: I1201 17:29:26.613102 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:29:26 crc kubenswrapper[4868]: I1201 17:29:26.650190 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:29:26 crc kubenswrapper[4868]: I1201 17:29:26.650264 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:29:26 crc kubenswrapper[4868]: I1201 17:29:26.678025 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:29:26 crc kubenswrapper[4868]: I1201 17:29:26.684136 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:29:26 crc kubenswrapper[4868]: I1201 17:29:26.696856 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:29:26 crc kubenswrapper[4868]: I1201 17:29:26.771836 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:29:26 crc kubenswrapper[4868]: I1201 17:29:26.781849 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:29:26 crc kubenswrapper[4868]: I1201 17:29:26.898257 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:29:26 crc kubenswrapper[4868]: I1201 17:29:26.898354 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:29:26 crc kubenswrapper[4868]: I1201 17:29:26.944438 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:29:27 crc kubenswrapper[4868]: I1201 17:29:27.767353 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:29:28 crc kubenswrapper[4868]: I1201 17:29:28.012258 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-grsv4"] Dec 01 17:29:28 crc kubenswrapper[4868]: I1201 17:29:28.141418 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:29:28 crc kubenswrapper[4868]: I1201 17:29:28.141495 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:29:28 crc kubenswrapper[4868]: I1201 17:29:28.207226 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:29:28 crc kubenswrapper[4868]: I1201 17:29:28.786368 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:29:28 crc kubenswrapper[4868]: I1201 17:29:28.970283 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:29:28 crc kubenswrapper[4868]: I1201 17:29:28.970334 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:29:29 crc kubenswrapper[4868]: I1201 17:29:29.011345 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:29:29 crc kubenswrapper[4868]: I1201 17:29:29.506014 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:29:29 crc kubenswrapper[4868]: I1201 17:29:29.506104 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:29:29 crc kubenswrapper[4868]: I1201 17:29:29.774042 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:29:29 crc kubenswrapper[4868]: I1201 17:29:29.853778 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:29:29 crc kubenswrapper[4868]: I1201 17:29:29.853845 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:29:30 crc kubenswrapper[4868]: I1201 17:29:30.546953 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2f8vp" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" containerName="registry-server" probeResult="failure" output=< Dec 01 17:29:30 crc kubenswrapper[4868]: timeout: failed to connect service ":50051" within 1s Dec 01 17:29:30 crc kubenswrapper[4868]: > Dec 01 17:29:30 crc kubenswrapper[4868]: I1201 17:29:30.889820 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pwdv6" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" containerName="registry-server" probeResult="failure" output=< Dec 01 17:29:30 crc kubenswrapper[4868]: timeout: failed to connect service ":50051" within 1s Dec 01 17:29:30 crc kubenswrapper[4868]: > Dec 01 17:29:31 crc kubenswrapper[4868]: I1201 17:29:31.324477 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hz4pc"] Dec 01 17:29:31 crc kubenswrapper[4868]: I1201 17:29:31.324823 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hz4pc" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" containerName="registry-server" containerID="cri-o://2cea147eb14fd0d23bcf87da73c105e7b8c83cea691afa775d90d6cc085c7348" gracePeriod=2 Dec 01 17:29:31 crc kubenswrapper[4868]: I1201 17:29:31.746834 4868 generic.go:334] "Generic (PLEG): container finished" podID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" containerID="2cea147eb14fd0d23bcf87da73c105e7b8c83cea691afa775d90d6cc085c7348" exitCode=0 Dec 01 17:29:31 crc kubenswrapper[4868]: I1201 17:29:31.746913 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hz4pc" event={"ID":"d7c0aa99-172b-41b7-80e5-2cbc6532cae1","Type":"ContainerDied","Data":"2cea147eb14fd0d23bcf87da73c105e7b8c83cea691afa775d90d6cc085c7348"} Dec 01 17:29:31 crc kubenswrapper[4868]: I1201 17:29:31.936467 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tjsn"] Dec 01 17:29:31 crc kubenswrapper[4868]: I1201 17:29:31.936805 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4tjsn" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" containerName="registry-server" containerID="cri-o://f085b3894b9b09f6dcc72340cb60a13cbc958ba4634f06de12a407cbb6d89274" gracePeriod=2 Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.213331 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.340855 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-utilities\") pod \"d7c0aa99-172b-41b7-80e5-2cbc6532cae1\" (UID: \"d7c0aa99-172b-41b7-80e5-2cbc6532cae1\") " Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.340977 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-catalog-content\") pod \"d7c0aa99-172b-41b7-80e5-2cbc6532cae1\" (UID: \"d7c0aa99-172b-41b7-80e5-2cbc6532cae1\") " Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.341110 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2jhx\" (UniqueName: \"kubernetes.io/projected/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-kube-api-access-n2jhx\") pod \"d7c0aa99-172b-41b7-80e5-2cbc6532cae1\" (UID: \"d7c0aa99-172b-41b7-80e5-2cbc6532cae1\") " Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.341905 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-utilities" (OuterVolumeSpecName: "utilities") pod "d7c0aa99-172b-41b7-80e5-2cbc6532cae1" (UID: "d7c0aa99-172b-41b7-80e5-2cbc6532cae1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.347278 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-kube-api-access-n2jhx" (OuterVolumeSpecName: "kube-api-access-n2jhx") pod "d7c0aa99-172b-41b7-80e5-2cbc6532cae1" (UID: "d7c0aa99-172b-41b7-80e5-2cbc6532cae1"). InnerVolumeSpecName "kube-api-access-n2jhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.443007 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2jhx\" (UniqueName: \"kubernetes.io/projected/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-kube-api-access-n2jhx\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.443051 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:32 crc kubenswrapper[4868]: E1201 17:29:32.690194 4868 file.go:109] "Unable to process watch event" err="can't process config file \"/etc/kubernetes/manifests/kube-apiserver-pod.yaml\": /etc/kubernetes/manifests/kube-apiserver-pod.yaml: couldn't parse as pod(Object 'Kind' is missing in 'null'), please check config file" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.690234 4868 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.690616 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30" gracePeriod=15 Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.690660 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152" gracePeriod=15 Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.690788 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e" gracePeriod=15 Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.690604 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7" gracePeriod=15 Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.690836 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9" gracePeriod=15 Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.702189 4868 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 17:29:32 crc kubenswrapper[4868]: E1201 17:29:32.704040 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.704095 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 17:29:32 crc kubenswrapper[4868]: E1201 17:29:32.704171 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.704184 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 17:29:32 crc kubenswrapper[4868]: E1201 17:29:32.704207 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.704219 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 17:29:32 crc kubenswrapper[4868]: E1201 17:29:32.704257 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" containerName="extract-content" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.704269 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" containerName="extract-content" Dec 01 17:29:32 crc kubenswrapper[4868]: E1201 17:29:32.704296 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.704306 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 17:29:32 crc kubenswrapper[4868]: E1201 17:29:32.704320 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" containerName="extract-utilities" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.704331 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" containerName="extract-utilities" Dec 01 17:29:32 crc kubenswrapper[4868]: E1201 17:29:32.704340 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.704350 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 17:29:32 crc kubenswrapper[4868]: E1201 17:29:32.704370 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" containerName="registry-server" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.704379 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" containerName="registry-server" Dec 01 17:29:32 crc kubenswrapper[4868]: E1201 17:29:32.704413 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da15d531-bc43-41f0-a568-a6a34308cfa3" containerName="pruner" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.704423 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="da15d531-bc43-41f0-a568-a6a34308cfa3" containerName="pruner" Dec 01 17:29:32 crc kubenswrapper[4868]: E1201 17:29:32.704444 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.704456 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.705114 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.705135 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.705144 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.705155 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" containerName="registry-server" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.705166 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="da15d531-bc43-41f0-a568-a6a34308cfa3" containerName="pruner" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.705178 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.705187 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 17:29:32 crc kubenswrapper[4868]: E1201 17:29:32.705616 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.705629 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.706060 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.711182 4868 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.727413 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.753147 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7c0aa99-172b-41b7-80e5-2cbc6532cae1" (UID: "d7c0aa99-172b-41b7-80e5-2cbc6532cae1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.770440 4868 generic.go:334] "Generic (PLEG): container finished" podID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" containerID="f085b3894b9b09f6dcc72340cb60a13cbc958ba4634f06de12a407cbb6d89274" exitCode=0 Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.770516 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tjsn" event={"ID":"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef","Type":"ContainerDied","Data":"f085b3894b9b09f6dcc72340cb60a13cbc958ba4634f06de12a407cbb6d89274"} Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.777964 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hz4pc" event={"ID":"d7c0aa99-172b-41b7-80e5-2cbc6532cae1","Type":"ContainerDied","Data":"f229bd27517b9ad6dde6edb651199086ad6f28f5e84ea204572f36d363476cb2"} Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.778046 4868 scope.go:117] "RemoveContainer" containerID="2cea147eb14fd0d23bcf87da73c105e7b8c83cea691afa775d90d6cc085c7348" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.778088 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hz4pc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.794108 4868 scope.go:117] "RemoveContainer" containerID="adf50d549d08c57b185d0a1fb9bd1dcbeb37dcdc1e20be21468322c3fa1e4f42" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.848624 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.848694 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.848742 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.848764 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.848804 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.849045 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.849141 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.849194 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.849356 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7c0aa99-172b-41b7-80e5-2cbc6532cae1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.872165 4868 scope.go:117] "RemoveContainer" containerID="959f721f35672a4e9bd5929a14dbede468422b3668484db48720077c48ce132e" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951290 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951343 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951368 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951450 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951469 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951493 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951487 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951555 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951512 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951599 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951632 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951639 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951657 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951690 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951703 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:32 crc kubenswrapper[4868]: I1201 17:29:32.951749 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.133732 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.256751 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-catalog-content\") pod \"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef\" (UID: \"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef\") " Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.256828 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlbmh\" (UniqueName: \"kubernetes.io/projected/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-kube-api-access-zlbmh\") pod \"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef\" (UID: \"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef\") " Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.256855 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-utilities\") pod \"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef\" (UID: \"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef\") " Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.258843 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-utilities" (OuterVolumeSpecName: "utilities") pod "14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" (UID: "14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.262613 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-kube-api-access-zlbmh" (OuterVolumeSpecName: "kube-api-access-zlbmh") pod "14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" (UID: "14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef"). InnerVolumeSpecName "kube-api-access-zlbmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.274450 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" (UID: "14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.358236 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.358280 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlbmh\" (UniqueName: \"kubernetes.io/projected/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-kube-api-access-zlbmh\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.358296 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.628629 4868 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.628708 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.788905 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.790371 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.791235 4868 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152" exitCode=0 Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.791324 4868 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9" exitCode=2 Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.803625 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tjsn" Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.804083 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tjsn" event={"ID":"14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef","Type":"ContainerDied","Data":"5847cf43309d5e24d03bc8e4dfc25f7b0751e057a427e42eeee97f646e0336f4"} Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.804308 4868 scope.go:117] "RemoveContainer" containerID="f085b3894b9b09f6dcc72340cb60a13cbc958ba4634f06de12a407cbb6d89274" Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.841797 4868 scope.go:117] "RemoveContainer" containerID="bc11c670e2c0d188d4cbe9eb7386bd5b695a99f3f44f959f0b45af2813f03135" Dec 01 17:29:33 crc kubenswrapper[4868]: I1201 17:29:33.862280 4868 scope.go:117] "RemoveContainer" containerID="2e60ea3c7ad9b897893213af7ecc2484b968c8f0f00e4b4e3be53c02bc483d65" Dec 01 17:29:34 crc kubenswrapper[4868]: I1201 17:29:34.827516 4868 generic.go:334] "Generic (PLEG): container finished" podID="ed9731d5-c84c-4e48-91c4-2e66fba896a6" containerID="4acf6c8a2136f0b65fce38151339a90aa6f197fc5aae62ff6aba2dae87234166" exitCode=0 Dec 01 17:29:34 crc kubenswrapper[4868]: I1201 17:29:34.828060 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ed9731d5-c84c-4e48-91c4-2e66fba896a6","Type":"ContainerDied","Data":"4acf6c8a2136f0b65fce38151339a90aa6f197fc5aae62ff6aba2dae87234166"} Dec 01 17:29:34 crc kubenswrapper[4868]: I1201 17:29:34.830923 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 17:29:34 crc kubenswrapper[4868]: I1201 17:29:34.833757 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 17:29:34 crc kubenswrapper[4868]: I1201 17:29:34.843293 4868 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30" exitCode=0 Dec 01 17:29:34 crc kubenswrapper[4868]: I1201 17:29:34.843351 4868 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e" exitCode=0 Dec 01 17:29:34 crc kubenswrapper[4868]: I1201 17:29:34.843479 4868 scope.go:117] "RemoveContainer" containerID="e9b4fe9757781f709d7ffc2585aa43a82d6d6fe617a1d3aca8539e5530485ebc" Dec 01 17:29:35 crc kubenswrapper[4868]: I1201 17:29:35.857189 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 17:29:35 crc kubenswrapper[4868]: I1201 17:29:35.859433 4868 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7" exitCode=0 Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.119578 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.120561 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.123649 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.210288 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.210357 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.210462 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ed9731d5-c84c-4e48-91c4-2e66fba896a6-var-lock\") pod \"ed9731d5-c84c-4e48-91c4-2e66fba896a6\" (UID: \"ed9731d5-c84c-4e48-91c4-2e66fba896a6\") " Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.210495 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed9731d5-c84c-4e48-91c4-2e66fba896a6-kubelet-dir\") pod \"ed9731d5-c84c-4e48-91c4-2e66fba896a6\" (UID: \"ed9731d5-c84c-4e48-91c4-2e66fba896a6\") " Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.210514 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.210578 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed9731d5-c84c-4e48-91c4-2e66fba896a6-kube-api-access\") pod \"ed9731d5-c84c-4e48-91c4-2e66fba896a6\" (UID: \"ed9731d5-c84c-4e48-91c4-2e66fba896a6\") " Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.210623 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.210687 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.210718 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed9731d5-c84c-4e48-91c4-2e66fba896a6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ed9731d5-c84c-4e48-91c4-2e66fba896a6" (UID: "ed9731d5-c84c-4e48-91c4-2e66fba896a6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.210731 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed9731d5-c84c-4e48-91c4-2e66fba896a6-var-lock" (OuterVolumeSpecName: "var-lock") pod "ed9731d5-c84c-4e48-91c4-2e66fba896a6" (UID: "ed9731d5-c84c-4e48-91c4-2e66fba896a6"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.210794 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.211664 4868 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ed9731d5-c84c-4e48-91c4-2e66fba896a6-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.211746 4868 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed9731d5-c84c-4e48-91c4-2e66fba896a6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.211771 4868 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.211790 4868 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.211843 4868 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.219637 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed9731d5-c84c-4e48-91c4-2e66fba896a6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ed9731d5-c84c-4e48-91c4-2e66fba896a6" (UID: "ed9731d5-c84c-4e48-91c4-2e66fba896a6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.313451 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed9731d5-c84c-4e48-91c4-2e66fba896a6-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.700524 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.865846 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ed9731d5-c84c-4e48-91c4-2e66fba896a6","Type":"ContainerDied","Data":"41f9965470ced10dd5b00144f999cee53bfeff593ae6a9f00955b7106016aef7"} Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.866912 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41f9965470ced10dd5b00144f999cee53bfeff593ae6a9f00955b7106016aef7" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.866003 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.869891 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.871140 4868 scope.go:117] "RemoveContainer" containerID="700ed14d0e657c1e6c15a6f6ccf45040db11ad5c411226342cc4e05ff1a64e30" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.871288 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.890158 4868 scope.go:117] "RemoveContainer" containerID="58664353e063f634ce9c94a65239f8cd71efbf59cb89ba114658e4d86177c152" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.908564 4868 scope.go:117] "RemoveContainer" containerID="f9e58c5ca7de2c817dc5257574b47957d569d78141f5761259b505de5a9cd16e" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.923128 4868 scope.go:117] "RemoveContainer" containerID="47c112b0e9f9946970b6164ad7ec526b0d4021d472ecf0605b7d0798697f58e9" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.952892 4868 scope.go:117] "RemoveContainer" containerID="7ff5053ae61c3b56b9e7d431c6a6e18e38e215c6b1bf88779aa0e58cc1fee1f7" Dec 01 17:29:36 crc kubenswrapper[4868]: I1201 17:29:36.970025 4868 scope.go:117] "RemoveContainer" containerID="696567c440c1493e4adad1c538205a712ebbfbf2477832b99d66b979eae119c9" Dec 01 17:29:37 crc kubenswrapper[4868]: I1201 17:29:37.731825 4868 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:37 crc kubenswrapper[4868]: I1201 17:29:37.733459 4868 status_manager.go:851] "Failed to get status for pod" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" pod="openshift-marketplace/certified-operators-c9tpp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c9tpp\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:37 crc kubenswrapper[4868]: I1201 17:29:37.734646 4868 status_manager.go:851] "Failed to get status for pod" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" pod="openshift-marketplace/community-operators-hz4pc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hz4pc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:37 crc kubenswrapper[4868]: I1201 17:29:37.735415 4868 status_manager.go:851] "Failed to get status for pod" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" pod="openshift-marketplace/redhat-marketplace-4tjsn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-4tjsn\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:37 crc kubenswrapper[4868]: I1201 17:29:37.736092 4868 status_manager.go:851] "Failed to get status for pod" podUID="ed9731d5-c84c-4e48-91c4-2e66fba896a6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:37 crc kubenswrapper[4868]: I1201 17:29:37.736745 4868 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:37 crc kubenswrapper[4868]: E1201 17:29:37.750467 4868 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.103:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:37 crc kubenswrapper[4868]: I1201 17:29:37.751772 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:37 crc kubenswrapper[4868]: W1201 17:29:37.805186 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-b925351be68f9a21483b482f86bdc6b0c93e5d61d6bcbc68fe1b9aec156c0be9 WatchSource:0}: Error finding container b925351be68f9a21483b482f86bdc6b0c93e5d61d6bcbc68fe1b9aec156c0be9: Status 404 returned error can't find the container with id b925351be68f9a21483b482f86bdc6b0c93e5d61d6bcbc68fe1b9aec156c0be9 Dec 01 17:29:37 crc kubenswrapper[4868]: E1201 17:29:37.813396 4868 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.103:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d279b7f62a8e6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 17:29:37.809688806 +0000 UTC m=+250.180799237,LastTimestamp:2025-12-01 17:29:37.809688806 +0000 UTC m=+250.180799237,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 17:29:37 crc kubenswrapper[4868]: I1201 17:29:37.883559 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b925351be68f9a21483b482f86bdc6b0c93e5d61d6bcbc68fe1b9aec156c0be9"} Dec 01 17:29:38 crc kubenswrapper[4868]: I1201 17:29:38.183475 4868 status_manager.go:851] "Failed to get status for pod" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" pod="openshift-marketplace/community-operators-hz4pc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hz4pc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:38 crc kubenswrapper[4868]: I1201 17:29:38.184208 4868 status_manager.go:851] "Failed to get status for pod" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" pod="openshift-marketplace/redhat-marketplace-4tjsn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-4tjsn\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:38 crc kubenswrapper[4868]: I1201 17:29:38.185251 4868 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:38 crc kubenswrapper[4868]: I1201 17:29:38.185448 4868 status_manager.go:851] "Failed to get status for pod" podUID="ed9731d5-c84c-4e48-91c4-2e66fba896a6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:38 crc kubenswrapper[4868]: I1201 17:29:38.185650 4868 status_manager.go:851] "Failed to get status for pod" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" pod="openshift-marketplace/certified-operators-c9tpp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c9tpp\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:38 crc kubenswrapper[4868]: I1201 17:29:38.190428 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 01 17:29:38 crc kubenswrapper[4868]: I1201 17:29:38.894334 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"179da301cbcf8aee82195bfae2f3fae1790e9cfcc06a0df19fefa0e4bc624ec9"} Dec 01 17:29:38 crc kubenswrapper[4868]: E1201 17:29:38.895483 4868 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.103:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:38 crc kubenswrapper[4868]: I1201 17:29:38.895522 4868 status_manager.go:851] "Failed to get status for pod" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" pod="openshift-marketplace/certified-operators-c9tpp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c9tpp\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:38 crc kubenswrapper[4868]: I1201 17:29:38.896324 4868 status_manager.go:851] "Failed to get status for pod" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" pod="openshift-marketplace/community-operators-hz4pc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hz4pc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:38 crc kubenswrapper[4868]: I1201 17:29:38.896695 4868 status_manager.go:851] "Failed to get status for pod" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" pod="openshift-marketplace/redhat-marketplace-4tjsn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-4tjsn\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:38 crc kubenswrapper[4868]: I1201 17:29:38.898280 4868 status_manager.go:851] "Failed to get status for pod" podUID="ed9731d5-c84c-4e48-91c4-2e66fba896a6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.563337 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.565129 4868 status_manager.go:851] "Failed to get status for pod" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" pod="openshift-marketplace/redhat-marketplace-4tjsn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-4tjsn\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.565555 4868 status_manager.go:851] "Failed to get status for pod" podUID="ed9731d5-c84c-4e48-91c4-2e66fba896a6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.565928 4868 status_manager.go:851] "Failed to get status for pod" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" pod="openshift-marketplace/redhat-operators-2f8vp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2f8vp\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.566291 4868 status_manager.go:851] "Failed to get status for pod" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" pod="openshift-marketplace/certified-operators-c9tpp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c9tpp\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.566977 4868 status_manager.go:851] "Failed to get status for pod" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" pod="openshift-marketplace/community-operators-hz4pc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hz4pc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.636667 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.637593 4868 status_manager.go:851] "Failed to get status for pod" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" pod="openshift-marketplace/redhat-marketplace-4tjsn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-4tjsn\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.638113 4868 status_manager.go:851] "Failed to get status for pod" podUID="ed9731d5-c84c-4e48-91c4-2e66fba896a6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.638534 4868 status_manager.go:851] "Failed to get status for pod" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" pod="openshift-marketplace/redhat-operators-2f8vp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2f8vp\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.638932 4868 status_manager.go:851] "Failed to get status for pod" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" pod="openshift-marketplace/certified-operators-c9tpp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c9tpp\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.639272 4868 status_manager.go:851] "Failed to get status for pod" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" pod="openshift-marketplace/community-operators-hz4pc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hz4pc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: E1201 17:29:39.920497 4868 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.103:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.922321 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.923370 4868 status_manager.go:851] "Failed to get status for pod" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" pod="openshift-marketplace/redhat-marketplace-4tjsn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-4tjsn\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.923908 4868 status_manager.go:851] "Failed to get status for pod" podUID="ed9731d5-c84c-4e48-91c4-2e66fba896a6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.924556 4868 status_manager.go:851] "Failed to get status for pod" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" pod="openshift-marketplace/redhat-operators-2f8vp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2f8vp\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.925258 4868 status_manager.go:851] "Failed to get status for pod" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" pod="openshift-marketplace/certified-operators-c9tpp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c9tpp\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.932647 4868 status_manager.go:851] "Failed to get status for pod" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" pod="openshift-marketplace/redhat-operators-pwdv6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-pwdv6\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.933277 4868 status_manager.go:851] "Failed to get status for pod" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" pod="openshift-marketplace/community-operators-hz4pc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hz4pc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.975299 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.976993 4868 status_manager.go:851] "Failed to get status for pod" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" pod="openshift-marketplace/redhat-marketplace-4tjsn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-4tjsn\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.978016 4868 status_manager.go:851] "Failed to get status for pod" podUID="ed9731d5-c84c-4e48-91c4-2e66fba896a6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.978542 4868 status_manager.go:851] "Failed to get status for pod" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" pod="openshift-marketplace/redhat-operators-2f8vp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2f8vp\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.979363 4868 status_manager.go:851] "Failed to get status for pod" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" pod="openshift-marketplace/certified-operators-c9tpp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c9tpp\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.979867 4868 status_manager.go:851] "Failed to get status for pod" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" pod="openshift-marketplace/redhat-operators-pwdv6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-pwdv6\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:39 crc kubenswrapper[4868]: I1201 17:29:39.980389 4868 status_manager.go:851] "Failed to get status for pod" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" pod="openshift-marketplace/community-operators-hz4pc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hz4pc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:41 crc kubenswrapper[4868]: E1201 17:29:41.067789 4868 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:41 crc kubenswrapper[4868]: E1201 17:29:41.068844 4868 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:41 crc kubenswrapper[4868]: E1201 17:29:41.069369 4868 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:41 crc kubenswrapper[4868]: E1201 17:29:41.069611 4868 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:41 crc kubenswrapper[4868]: E1201 17:29:41.069892 4868 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:41 crc kubenswrapper[4868]: I1201 17:29:41.070152 4868 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 01 17:29:41 crc kubenswrapper[4868]: E1201 17:29:41.070362 4868 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" interval="200ms" Dec 01 17:29:41 crc kubenswrapper[4868]: E1201 17:29:41.271977 4868 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" interval="400ms" Dec 01 17:29:41 crc kubenswrapper[4868]: E1201 17:29:41.673731 4868 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" interval="800ms" Dec 01 17:29:42 crc kubenswrapper[4868]: E1201 17:29:42.475625 4868 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" interval="1.6s" Dec 01 17:29:44 crc kubenswrapper[4868]: E1201 17:29:44.077230 4868 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.103:6443: connect: connection refused" interval="3.2s" Dec 01 17:29:45 crc kubenswrapper[4868]: E1201 17:29:45.156295 4868 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.103:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d279b7f62a8e6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 17:29:37.809688806 +0000 UTC m=+250.180799237,LastTimestamp:2025-12-01 17:29:37.809688806 +0000 UTC m=+250.180799237,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.171760 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.172976 4868 status_manager.go:851] "Failed to get status for pod" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" pod="openshift-marketplace/certified-operators-c9tpp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c9tpp\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.173437 4868 status_manager.go:851] "Failed to get status for pod" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" pod="openshift-marketplace/redhat-operators-pwdv6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-pwdv6\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.174088 4868 status_manager.go:851] "Failed to get status for pod" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" pod="openshift-marketplace/community-operators-hz4pc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hz4pc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.174544 4868 status_manager.go:851] "Failed to get status for pod" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" pod="openshift-marketplace/redhat-marketplace-4tjsn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-4tjsn\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.174816 4868 status_manager.go:851] "Failed to get status for pod" podUID="ed9731d5-c84c-4e48-91c4-2e66fba896a6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.175133 4868 status_manager.go:851] "Failed to get status for pod" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" pod="openshift-marketplace/redhat-operators-2f8vp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2f8vp\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.188576 4868 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7831733f-2e93-4000-95c1-d5cd30cbaeee" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.188608 4868 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7831733f-2e93-4000-95c1-d5cd30cbaeee" Dec 01 17:29:45 crc kubenswrapper[4868]: E1201 17:29:45.189208 4868 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.103:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.190187 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.950813 4868 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="3955aad205670bdd28027342d18a9179b5c93fbdefcfa47164e7bfaacf5d1ee2" exitCode=0 Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.951120 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"3955aad205670bdd28027342d18a9179b5c93fbdefcfa47164e7bfaacf5d1ee2"} Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.951205 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"12699114046a4b7d21c7f51ee0f4f3a23576d4a98020e162a624b25208dd3169"} Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.951734 4868 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7831733f-2e93-4000-95c1-d5cd30cbaeee" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.951768 4868 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7831733f-2e93-4000-95c1-d5cd30cbaeee" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.952138 4868 status_manager.go:851] "Failed to get status for pod" podUID="ed9731d5-c84c-4e48-91c4-2e66fba896a6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.952463 4868 status_manager.go:851] "Failed to get status for pod" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" pod="openshift-marketplace/redhat-operators-2f8vp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2f8vp\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:45 crc kubenswrapper[4868]: E1201 17:29:45.952498 4868 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.103:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.952686 4868 status_manager.go:851] "Failed to get status for pod" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" pod="openshift-marketplace/redhat-operators-pwdv6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-pwdv6\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.952894 4868 status_manager.go:851] "Failed to get status for pod" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" pod="openshift-marketplace/certified-operators-c9tpp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-c9tpp\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.953120 4868 status_manager.go:851] "Failed to get status for pod" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" pod="openshift-marketplace/community-operators-hz4pc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hz4pc\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:45 crc kubenswrapper[4868]: I1201 17:29:45.953323 4868 status_manager.go:851] "Failed to get status for pod" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" pod="openshift-marketplace/redhat-marketplace-4tjsn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-4tjsn\": dial tcp 38.102.83.103:6443: connect: connection refused" Dec 01 17:29:46 crc kubenswrapper[4868]: I1201 17:29:46.961574 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 17:29:46 crc kubenswrapper[4868]: I1201 17:29:46.962691 4868 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6" exitCode=1 Dec 01 17:29:46 crc kubenswrapper[4868]: I1201 17:29:46.962797 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6"} Dec 01 17:29:46 crc kubenswrapper[4868]: I1201 17:29:46.963490 4868 scope.go:117] "RemoveContainer" containerID="301cd03f83d471d05bd41ea8a0f1daa5e49e0859f70e423a2f7d2aa88101f1a6" Dec 01 17:29:46 crc kubenswrapper[4868]: I1201 17:29:46.969483 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"bcbb88e88e7ebec4ba146496ce0a31486b669c79d9edc28fbedeb451ae98fe83"} Dec 01 17:29:46 crc kubenswrapper[4868]: I1201 17:29:46.969608 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a46aa5aca9221d220feea0dab82644bbea3c1efa94e1b6bdf53de3b7dcae48d6"} Dec 01 17:29:46 crc kubenswrapper[4868]: I1201 17:29:46.969680 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"02db4b5c2a790525814cdf7c850bb5309e0e29de80d7304af967fe369cba3158"} Dec 01 17:29:46 crc kubenswrapper[4868]: I1201 17:29:46.969747 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"63cf25019e5931cb5042cd16b1188e534a868176cd3b10b95e420b5e56f9d47d"} Dec 01 17:29:47 crc kubenswrapper[4868]: I1201 17:29:47.979104 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ca6afb4d0ea259869105867f323d30b3c9896b0f6618300d313956ad727130fa"} Dec 01 17:29:47 crc kubenswrapper[4868]: I1201 17:29:47.979469 4868 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7831733f-2e93-4000-95c1-d5cd30cbaeee" Dec 01 17:29:47 crc kubenswrapper[4868]: I1201 17:29:47.979490 4868 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7831733f-2e93-4000-95c1-d5cd30cbaeee" Dec 01 17:29:47 crc kubenswrapper[4868]: I1201 17:29:47.979834 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:47 crc kubenswrapper[4868]: I1201 17:29:47.983993 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 17:29:47 crc kubenswrapper[4868]: I1201 17:29:47.984040 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"491d0b23686dbc0ae5bd6eaed944b6c2d656ba33f8cd72ba5fa3581008bc5400"} Dec 01 17:29:48 crc kubenswrapper[4868]: I1201 17:29:48.249893 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:29:48 crc kubenswrapper[4868]: I1201 17:29:48.250456 4868 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 01 17:29:48 crc kubenswrapper[4868]: I1201 17:29:48.250620 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 01 17:29:48 crc kubenswrapper[4868]: I1201 17:29:48.771640 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:29:50 crc kubenswrapper[4868]: I1201 17:29:50.191075 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:50 crc kubenswrapper[4868]: I1201 17:29:50.191414 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:50 crc kubenswrapper[4868]: I1201 17:29:50.196281 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:52 crc kubenswrapper[4868]: I1201 17:29:52.991541 4868 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:29:52 crc kubenswrapper[4868]: I1201 17:29:52.996264 4868 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7831733f-2e93-4000-95c1-d5cd30cbaeee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:29:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:29:45Z\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:29:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T17:29:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63cf25019e5931cb5042cd16b1188e534a868176cd3b10b95e420b5e56f9d47d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:29:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a46aa5aca9221d220feea0dab82644bbea3c1efa94e1b6bdf53de3b7dcae48d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:29:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02db4b5c2a790525814cdf7c850bb5309e0e29de80d7304af967fe369cba3158\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:29:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca6afb4d0ea259869105867f323d30b3c9896b0f6618300d313956ad727130fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:29:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcbb88e88e7ebec4ba146496ce0a31486b669c79d9edc28fbedeb451ae98fe83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T17:29:46Z\\\"}}}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3955aad205670bdd28027342d18a9179b5c93fbdefcfa47164e7bfaacf5d1ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3955aad205670bdd28027342d18a9179b5c93fbdefcfa47164e7bfaacf5d1ee2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T17:29:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T17:29:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}]}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": pods \"kube-apiserver-crc\" not found" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.042513 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" podUID="6749e2f1-9c2c-41ce-8487-44f1a7ff7576" containerName="oauth-openshift" containerID="cri-o://5599db4f6b49491c9ecd619e10ef3425d7732c6c60af4c98188127a421e6ddcf" gracePeriod=15 Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.384727 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.390975 4868 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d808f4ee-5c63-4fca-b670-a04b8c0010c8" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.481323 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-trusted-ca-bundle\") pod \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.481388 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-service-ca\") pod \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.481438 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-audit-policies\") pod \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.481463 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-error\") pod \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.481481 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-router-certs\") pod \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.481522 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-session\") pod \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.481540 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-idp-0-file-data\") pod \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.481560 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-provider-selection\") pod \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.481582 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-audit-dir\") pod \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.481611 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppv8z\" (UniqueName: \"kubernetes.io/projected/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-kube-api-access-ppv8z\") pod \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.481659 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-serving-cert\") pod \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.481677 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-ocp-branding-template\") pod \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.481702 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-cliconfig\") pod \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.481773 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "6749e2f1-9c2c-41ce-8487-44f1a7ff7576" (UID: "6749e2f1-9c2c-41ce-8487-44f1a7ff7576"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.482276 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "6749e2f1-9c2c-41ce-8487-44f1a7ff7576" (UID: "6749e2f1-9c2c-41ce-8487-44f1a7ff7576"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.482294 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "6749e2f1-9c2c-41ce-8487-44f1a7ff7576" (UID: "6749e2f1-9c2c-41ce-8487-44f1a7ff7576"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.482362 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "6749e2f1-9c2c-41ce-8487-44f1a7ff7576" (UID: "6749e2f1-9c2c-41ce-8487-44f1a7ff7576"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.482547 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-login\") pod \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\" (UID: \"6749e2f1-9c2c-41ce-8487-44f1a7ff7576\") " Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.482932 4868 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.483036 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "6749e2f1-9c2c-41ce-8487-44f1a7ff7576" (UID: "6749e2f1-9c2c-41ce-8487-44f1a7ff7576"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.492091 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-kube-api-access-ppv8z" (OuterVolumeSpecName: "kube-api-access-ppv8z") pod "6749e2f1-9c2c-41ce-8487-44f1a7ff7576" (UID: "6749e2f1-9c2c-41ce-8487-44f1a7ff7576"). InnerVolumeSpecName "kube-api-access-ppv8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.492135 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "6749e2f1-9c2c-41ce-8487-44f1a7ff7576" (UID: "6749e2f1-9c2c-41ce-8487-44f1a7ff7576"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.494711 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "6749e2f1-9c2c-41ce-8487-44f1a7ff7576" (UID: "6749e2f1-9c2c-41ce-8487-44f1a7ff7576"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.495009 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "6749e2f1-9c2c-41ce-8487-44f1a7ff7576" (UID: "6749e2f1-9c2c-41ce-8487-44f1a7ff7576"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.496403 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "6749e2f1-9c2c-41ce-8487-44f1a7ff7576" (UID: "6749e2f1-9c2c-41ce-8487-44f1a7ff7576"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.506863 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "6749e2f1-9c2c-41ce-8487-44f1a7ff7576" (UID: "6749e2f1-9c2c-41ce-8487-44f1a7ff7576"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.507322 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "6749e2f1-9c2c-41ce-8487-44f1a7ff7576" (UID: "6749e2f1-9c2c-41ce-8487-44f1a7ff7576"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.507788 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "6749e2f1-9c2c-41ce-8487-44f1a7ff7576" (UID: "6749e2f1-9c2c-41ce-8487-44f1a7ff7576"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.508066 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "6749e2f1-9c2c-41ce-8487-44f1a7ff7576" (UID: "6749e2f1-9c2c-41ce-8487-44f1a7ff7576"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.584148 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.584207 4868 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.584223 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.584240 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.584283 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.584296 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.584311 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.584326 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppv8z\" (UniqueName: \"kubernetes.io/projected/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-kube-api-access-ppv8z\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.584339 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.584351 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.584363 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.584376 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:53 crc kubenswrapper[4868]: I1201 17:29:53.584389 4868 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6749e2f1-9c2c-41ce-8487-44f1a7ff7576-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:29:54 crc kubenswrapper[4868]: I1201 17:29:54.024398 4868 generic.go:334] "Generic (PLEG): container finished" podID="6749e2f1-9c2c-41ce-8487-44f1a7ff7576" containerID="5599db4f6b49491c9ecd619e10ef3425d7732c6c60af4c98188127a421e6ddcf" exitCode=0 Dec 01 17:29:54 crc kubenswrapper[4868]: I1201 17:29:54.024508 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" Dec 01 17:29:54 crc kubenswrapper[4868]: I1201 17:29:54.024544 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" event={"ID":"6749e2f1-9c2c-41ce-8487-44f1a7ff7576","Type":"ContainerDied","Data":"5599db4f6b49491c9ecd619e10ef3425d7732c6c60af4c98188127a421e6ddcf"} Dec 01 17:29:54 crc kubenswrapper[4868]: I1201 17:29:54.024703 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-grsv4" event={"ID":"6749e2f1-9c2c-41ce-8487-44f1a7ff7576","Type":"ContainerDied","Data":"4de7f02a896425ddc2f2fd860b389e7f856e98f308a8766b06565a35ceedcabc"} Dec 01 17:29:54 crc kubenswrapper[4868]: I1201 17:29:54.024767 4868 scope.go:117] "RemoveContainer" containerID="5599db4f6b49491c9ecd619e10ef3425d7732c6c60af4c98188127a421e6ddcf" Dec 01 17:29:54 crc kubenswrapper[4868]: I1201 17:29:54.025169 4868 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7831733f-2e93-4000-95c1-d5cd30cbaeee" Dec 01 17:29:54 crc kubenswrapper[4868]: I1201 17:29:54.025230 4868 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7831733f-2e93-4000-95c1-d5cd30cbaeee" Dec 01 17:29:54 crc kubenswrapper[4868]: I1201 17:29:54.042815 4868 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d808f4ee-5c63-4fca-b670-a04b8c0010c8" Dec 01 17:29:54 crc kubenswrapper[4868]: I1201 17:29:54.055558 4868 scope.go:117] "RemoveContainer" containerID="5599db4f6b49491c9ecd619e10ef3425d7732c6c60af4c98188127a421e6ddcf" Dec 01 17:29:54 crc kubenswrapper[4868]: E1201 17:29:54.056162 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5599db4f6b49491c9ecd619e10ef3425d7732c6c60af4c98188127a421e6ddcf\": container with ID starting with 5599db4f6b49491c9ecd619e10ef3425d7732c6c60af4c98188127a421e6ddcf not found: ID does not exist" containerID="5599db4f6b49491c9ecd619e10ef3425d7732c6c60af4c98188127a421e6ddcf" Dec 01 17:29:54 crc kubenswrapper[4868]: I1201 17:29:54.056199 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5599db4f6b49491c9ecd619e10ef3425d7732c6c60af4c98188127a421e6ddcf"} err="failed to get container status \"5599db4f6b49491c9ecd619e10ef3425d7732c6c60af4c98188127a421e6ddcf\": rpc error: code = NotFound desc = could not find container \"5599db4f6b49491c9ecd619e10ef3425d7732c6c60af4c98188127a421e6ddcf\": container with ID starting with 5599db4f6b49491c9ecd619e10ef3425d7732c6c60af4c98188127a421e6ddcf not found: ID does not exist" Dec 01 17:29:58 crc kubenswrapper[4868]: I1201 17:29:58.253394 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:29:58 crc kubenswrapper[4868]: I1201 17:29:58.257772 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 17:30:02 crc kubenswrapper[4868]: I1201 17:30:02.635704 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 17:30:02 crc kubenswrapper[4868]: I1201 17:30:02.694685 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 17:30:03 crc kubenswrapper[4868]: I1201 17:30:03.148028 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 17:30:04 crc kubenswrapper[4868]: I1201 17:30:04.019180 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 17:30:04 crc kubenswrapper[4868]: I1201 17:30:04.087063 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 17:30:04 crc kubenswrapper[4868]: I1201 17:30:04.235761 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 17:30:04 crc kubenswrapper[4868]: I1201 17:30:04.270005 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 17:30:04 crc kubenswrapper[4868]: I1201 17:30:04.426106 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 17:30:04 crc kubenswrapper[4868]: I1201 17:30:04.734992 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 17:30:04 crc kubenswrapper[4868]: I1201 17:30:04.794241 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 17:30:04 crc kubenswrapper[4868]: I1201 17:30:04.848475 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 17:30:04 crc kubenswrapper[4868]: I1201 17:30:04.910218 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 17:30:04 crc kubenswrapper[4868]: I1201 17:30:04.911021 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 17:30:05 crc kubenswrapper[4868]: I1201 17:30:05.050959 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 17:30:05 crc kubenswrapper[4868]: I1201 17:30:05.062568 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 17:30:05 crc kubenswrapper[4868]: I1201 17:30:05.085589 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 17:30:05 crc kubenswrapper[4868]: I1201 17:30:05.264883 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 17:30:05 crc kubenswrapper[4868]: I1201 17:30:05.300306 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 17:30:05 crc kubenswrapper[4868]: I1201 17:30:05.510995 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 17:30:05 crc kubenswrapper[4868]: I1201 17:30:05.728449 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 17:30:05 crc kubenswrapper[4868]: I1201 17:30:05.741404 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 17:30:05 crc kubenswrapper[4868]: I1201 17:30:05.742212 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 17:30:05 crc kubenswrapper[4868]: I1201 17:30:05.787715 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 17:30:05 crc kubenswrapper[4868]: I1201 17:30:05.940291 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 17:30:05 crc kubenswrapper[4868]: I1201 17:30:05.952816 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 17:30:05 crc kubenswrapper[4868]: I1201 17:30:05.975040 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 17:30:06 crc kubenswrapper[4868]: I1201 17:30:06.160071 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 17:30:06 crc kubenswrapper[4868]: I1201 17:30:06.185008 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 17:30:06 crc kubenswrapper[4868]: I1201 17:30:06.267968 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 17:30:06 crc kubenswrapper[4868]: I1201 17:30:06.371674 4868 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 17:30:06 crc kubenswrapper[4868]: I1201 17:30:06.663235 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 17:30:06 crc kubenswrapper[4868]: I1201 17:30:06.727518 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 17:30:06 crc kubenswrapper[4868]: I1201 17:30:06.752768 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 17:30:06 crc kubenswrapper[4868]: I1201 17:30:06.753964 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 17:30:06 crc kubenswrapper[4868]: I1201 17:30:06.936010 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 17:30:06 crc kubenswrapper[4868]: I1201 17:30:06.999478 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.061690 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.329144 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.424500 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.465149 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.468065 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.498912 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.532389 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.619015 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.641489 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.692159 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.707150 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.783775 4868 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.810591 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.842977 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.859716 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.861921 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.900025 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.911869 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.944163 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 17:30:07 crc kubenswrapper[4868]: I1201 17:30:07.975476 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 17:30:08 crc kubenswrapper[4868]: I1201 17:30:08.042107 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 17:30:08 crc kubenswrapper[4868]: I1201 17:30:08.067084 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 17:30:08 crc kubenswrapper[4868]: I1201 17:30:08.105067 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 17:30:08 crc kubenswrapper[4868]: I1201 17:30:08.111457 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 17:30:08 crc kubenswrapper[4868]: I1201 17:30:08.133889 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 17:30:08 crc kubenswrapper[4868]: I1201 17:30:08.158663 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 17:30:08 crc kubenswrapper[4868]: I1201 17:30:08.266461 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 17:30:08 crc kubenswrapper[4868]: I1201 17:30:08.402466 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 17:30:08 crc kubenswrapper[4868]: I1201 17:30:08.493781 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 17:30:08 crc kubenswrapper[4868]: I1201 17:30:08.686898 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 17:30:08 crc kubenswrapper[4868]: I1201 17:30:08.759260 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 17:30:08 crc kubenswrapper[4868]: I1201 17:30:08.904603 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 17:30:08 crc kubenswrapper[4868]: I1201 17:30:08.936933 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 17:30:08 crc kubenswrapper[4868]: I1201 17:30:08.961669 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 17:30:08 crc kubenswrapper[4868]: I1201 17:30:08.967245 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.014747 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.073644 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.118447 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.125012 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.264441 4868 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.304630 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.370054 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.405208 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.409661 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.573663 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.595811 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.597246 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.623826 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.742121 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.751463 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.807613 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.897654 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.908478 4868 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.914269 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hz4pc","openshift-authentication/oauth-openshift-558db77b4-grsv4","openshift-kube-apiserver/kube-apiserver-crc","openshift-marketplace/redhat-marketplace-4tjsn"] Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.914346 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s"] Dec 01 17:30:09 crc kubenswrapper[4868]: E1201 17:30:09.914627 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" containerName="extract-content" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.914656 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" containerName="extract-content" Dec 01 17:30:09 crc kubenswrapper[4868]: E1201 17:30:09.915326 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed9731d5-c84c-4e48-91c4-2e66fba896a6" containerName="installer" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.915345 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed9731d5-c84c-4e48-91c4-2e66fba896a6" containerName="installer" Dec 01 17:30:09 crc kubenswrapper[4868]: E1201 17:30:09.915371 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" containerName="extract-utilities" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.915380 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" containerName="extract-utilities" Dec 01 17:30:09 crc kubenswrapper[4868]: E1201 17:30:09.915390 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6749e2f1-9c2c-41ce-8487-44f1a7ff7576" containerName="oauth-openshift" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.915399 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="6749e2f1-9c2c-41ce-8487-44f1a7ff7576" containerName="oauth-openshift" Dec 01 17:30:09 crc kubenswrapper[4868]: E1201 17:30:09.915410 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" containerName="registry-server" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.915420 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" containerName="registry-server" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.915552 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="6749e2f1-9c2c-41ce-8487-44f1a7ff7576" containerName="oauth-openshift" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.915573 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" containerName="registry-server" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.915593 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed9731d5-c84c-4e48-91c4-2e66fba896a6" containerName="installer" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.916436 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.919007 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.919075 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.920191 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.939417 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.939393381 podStartE2EDuration="16.939393381s" podCreationTimestamp="2025-12-01 17:29:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:30:09.937048373 +0000 UTC m=+282.308158794" watchObservedRunningTime="2025-12-01 17:30:09.939393381 +0000 UTC m=+282.310503792" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.941688 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.975715 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 17:30:09 crc kubenswrapper[4868]: I1201 17:30:09.995116 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.038913 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.048090 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.048235 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.062859 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee6b0713-849b-4108-8fa0-53739ec95458-config-volume\") pod \"collect-profiles-29410170-4w22s\" (UID: \"ee6b0713-849b-4108-8fa0-53739ec95458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.063040 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrlgx\" (UniqueName: \"kubernetes.io/projected/ee6b0713-849b-4108-8fa0-53739ec95458-kube-api-access-nrlgx\") pod \"collect-profiles-29410170-4w22s\" (UID: \"ee6b0713-849b-4108-8fa0-53739ec95458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.063070 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee6b0713-849b-4108-8fa0-53739ec95458-secret-volume\") pod \"collect-profiles-29410170-4w22s\" (UID: \"ee6b0713-849b-4108-8fa0-53739ec95458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.126891 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.130623 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.132311 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.165482 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrlgx\" (UniqueName: \"kubernetes.io/projected/ee6b0713-849b-4108-8fa0-53739ec95458-kube-api-access-nrlgx\") pod \"collect-profiles-29410170-4w22s\" (UID: \"ee6b0713-849b-4108-8fa0-53739ec95458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.165586 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee6b0713-849b-4108-8fa0-53739ec95458-secret-volume\") pod \"collect-profiles-29410170-4w22s\" (UID: \"ee6b0713-849b-4108-8fa0-53739ec95458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.165634 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee6b0713-849b-4108-8fa0-53739ec95458-config-volume\") pod \"collect-profiles-29410170-4w22s\" (UID: \"ee6b0713-849b-4108-8fa0-53739ec95458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.167683 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee6b0713-849b-4108-8fa0-53739ec95458-config-volume\") pod \"collect-profiles-29410170-4w22s\" (UID: \"ee6b0713-849b-4108-8fa0-53739ec95458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.176285 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee6b0713-849b-4108-8fa0-53739ec95458-secret-volume\") pod \"collect-profiles-29410170-4w22s\" (UID: \"ee6b0713-849b-4108-8fa0-53739ec95458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.184479 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrlgx\" (UniqueName: \"kubernetes.io/projected/ee6b0713-849b-4108-8fa0-53739ec95458-kube-api-access-nrlgx\") pod \"collect-profiles-29410170-4w22s\" (UID: \"ee6b0713-849b-4108-8fa0-53739ec95458\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.187886 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef" path="/var/lib/kubelet/pods/14bae962-c4a6-4f4a-b6dd-435d2ddeb9ef/volumes" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.189751 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6749e2f1-9c2c-41ce-8487-44f1a7ff7576" path="/var/lib/kubelet/pods/6749e2f1-9c2c-41ce-8487-44f1a7ff7576/volumes" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.191045 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7c0aa99-172b-41b7-80e5-2cbc6532cae1" path="/var/lib/kubelet/pods/d7c0aa99-172b-41b7-80e5-2cbc6532cae1/volumes" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.214594 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.218370 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.237961 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.273136 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.286861 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.402889 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.408648 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.575983 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.577244 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.635544 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.665486 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.713180 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.747813 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 17:30:10 crc kubenswrapper[4868]: I1201 17:30:10.756868 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 17:30:11 crc kubenswrapper[4868]: I1201 17:30:11.035278 4868 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 17:30:11 crc kubenswrapper[4868]: I1201 17:30:11.166692 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 17:30:11 crc kubenswrapper[4868]: I1201 17:30:11.169435 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 17:30:11 crc kubenswrapper[4868]: I1201 17:30:11.223840 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 17:30:11 crc kubenswrapper[4868]: I1201 17:30:11.254191 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 17:30:11 crc kubenswrapper[4868]: I1201 17:30:11.296074 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 17:30:11 crc kubenswrapper[4868]: I1201 17:30:11.349394 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 17:30:11 crc kubenswrapper[4868]: I1201 17:30:11.441159 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 17:30:11 crc kubenswrapper[4868]: I1201 17:30:11.558957 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 17:30:11 crc kubenswrapper[4868]: I1201 17:30:11.629505 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 17:30:11 crc kubenswrapper[4868]: I1201 17:30:11.634476 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 17:30:11 crc kubenswrapper[4868]: I1201 17:30:11.758480 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 17:30:11 crc kubenswrapper[4868]: I1201 17:30:11.758493 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 17:30:11 crc kubenswrapper[4868]: I1201 17:30:11.770021 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 17:30:11 crc kubenswrapper[4868]: I1201 17:30:11.923587 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 17:30:12 crc kubenswrapper[4868]: I1201 17:30:12.006874 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 17:30:12 crc kubenswrapper[4868]: I1201 17:30:12.168004 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 17:30:12 crc kubenswrapper[4868]: I1201 17:30:12.209234 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 17:30:12 crc kubenswrapper[4868]: I1201 17:30:12.378042 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 17:30:12 crc kubenswrapper[4868]: I1201 17:30:12.451580 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 17:30:12 crc kubenswrapper[4868]: I1201 17:30:12.470371 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 17:30:12 crc kubenswrapper[4868]: I1201 17:30:12.515918 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 17:30:12 crc kubenswrapper[4868]: I1201 17:30:12.546030 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 17:30:12 crc kubenswrapper[4868]: I1201 17:30:12.570484 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 17:30:12 crc kubenswrapper[4868]: I1201 17:30:12.788476 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 17:30:12 crc kubenswrapper[4868]: I1201 17:30:12.797702 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 17:30:12 crc kubenswrapper[4868]: I1201 17:30:12.810134 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 17:30:12 crc kubenswrapper[4868]: I1201 17:30:12.937003 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 17:30:13 crc kubenswrapper[4868]: I1201 17:30:13.041001 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 17:30:13 crc kubenswrapper[4868]: I1201 17:30:13.051927 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 17:30:13 crc kubenswrapper[4868]: I1201 17:30:13.099510 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 17:30:13 crc kubenswrapper[4868]: I1201 17:30:13.164779 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 17:30:13 crc kubenswrapper[4868]: I1201 17:30:13.175172 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 17:30:13 crc kubenswrapper[4868]: I1201 17:30:13.232398 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 17:30:13 crc kubenswrapper[4868]: I1201 17:30:13.272307 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 17:30:13 crc kubenswrapper[4868]: I1201 17:30:13.309876 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s"] Dec 01 17:30:13 crc kubenswrapper[4868]: I1201 17:30:13.416381 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 17:30:13 crc kubenswrapper[4868]: I1201 17:30:13.608955 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s"] Dec 01 17:30:13 crc kubenswrapper[4868]: I1201 17:30:13.663592 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 17:30:13 crc kubenswrapper[4868]: I1201 17:30:13.683813 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.018546 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.054746 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.073298 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.073406 4868 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.151775 4868 generic.go:334] "Generic (PLEG): container finished" podID="ee6b0713-849b-4108-8fa0-53739ec95458" containerID="8f5fa2feac10e3963105ad4104d3c6ac4c849489a67233fd901c985ec658978c" exitCode=0 Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.151850 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" event={"ID":"ee6b0713-849b-4108-8fa0-53739ec95458","Type":"ContainerDied","Data":"8f5fa2feac10e3963105ad4104d3c6ac4c849489a67233fd901c985ec658978c"} Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.151894 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" event={"ID":"ee6b0713-849b-4108-8fa0-53739ec95458","Type":"ContainerStarted","Data":"93b365df74e3827364246978c264e3d204e2bd1996d2aefbc8d5def5272341e2"} Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.154602 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.165989 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.291755 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.322455 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.324042 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.338484 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.371901 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.375646 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.375982 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.389349 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.406043 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.537321 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.568029 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.577602 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.605318 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.619031 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.718191 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.823435 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.926216 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 17:30:14 crc kubenswrapper[4868]: I1201 17:30:14.928833 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.140846 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.183126 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.232416 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.237127 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.243571 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.267766 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.307122 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.462569 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.490317 4868 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.490611 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://179da301cbcf8aee82195bfae2f3fae1790e9cfcc06a0df19fefa0e4bc624ec9" gracePeriod=5 Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.492533 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.546177 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee6b0713-849b-4108-8fa0-53739ec95458-config-volume\") pod \"ee6b0713-849b-4108-8fa0-53739ec95458\" (UID: \"ee6b0713-849b-4108-8fa0-53739ec95458\") " Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.546313 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee6b0713-849b-4108-8fa0-53739ec95458-secret-volume\") pod \"ee6b0713-849b-4108-8fa0-53739ec95458\" (UID: \"ee6b0713-849b-4108-8fa0-53739ec95458\") " Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.546386 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrlgx\" (UniqueName: \"kubernetes.io/projected/ee6b0713-849b-4108-8fa0-53739ec95458-kube-api-access-nrlgx\") pod \"ee6b0713-849b-4108-8fa0-53739ec95458\" (UID: \"ee6b0713-849b-4108-8fa0-53739ec95458\") " Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.547187 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee6b0713-849b-4108-8fa0-53739ec95458-config-volume" (OuterVolumeSpecName: "config-volume") pod "ee6b0713-849b-4108-8fa0-53739ec95458" (UID: "ee6b0713-849b-4108-8fa0-53739ec95458"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.553751 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee6b0713-849b-4108-8fa0-53739ec95458-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ee6b0713-849b-4108-8fa0-53739ec95458" (UID: "ee6b0713-849b-4108-8fa0-53739ec95458"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.556146 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee6b0713-849b-4108-8fa0-53739ec95458-kube-api-access-nrlgx" (OuterVolumeSpecName: "kube-api-access-nrlgx") pod "ee6b0713-849b-4108-8fa0-53739ec95458" (UID: "ee6b0713-849b-4108-8fa0-53739ec95458"). InnerVolumeSpecName "kube-api-access-nrlgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.629844 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.647905 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrlgx\" (UniqueName: \"kubernetes.io/projected/ee6b0713-849b-4108-8fa0-53739ec95458-kube-api-access-nrlgx\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.647956 4868 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee6b0713-849b-4108-8fa0-53739ec95458-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.647971 4868 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee6b0713-849b-4108-8fa0-53739ec95458-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.662870 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.716144 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.764645 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.797284 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.891715 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.936442 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.945338 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.947361 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-ccc74cc7-zj2l6"] Dec 01 17:30:15 crc kubenswrapper[4868]: E1201 17:30:15.947633 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee6b0713-849b-4108-8fa0-53739ec95458" containerName="collect-profiles" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.947652 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee6b0713-849b-4108-8fa0-53739ec95458" containerName="collect-profiles" Dec 01 17:30:15 crc kubenswrapper[4868]: E1201 17:30:15.947674 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.947681 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.947774 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.947784 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee6b0713-849b-4108-8fa0-53739ec95458" containerName="collect-profiles" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.948242 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.951439 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.952428 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.953296 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.953348 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.953516 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.953933 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.954652 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.955915 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.956014 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.956099 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.961388 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.961963 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.968479 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.975115 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 17:30:15 crc kubenswrapper[4868]: I1201 17:30:15.976162 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-ccc74cc7-zj2l6"] Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.018573 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.025268 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.027126 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.048306 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.054552 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-user-template-login\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.055435 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ct67\" (UniqueName: \"kubernetes.io/projected/f50f2600-78a9-4175-b0af-c5a18563e39d-kube-api-access-2ct67\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.055544 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f50f2600-78a9-4175-b0af-c5a18563e39d-audit-dir\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.055656 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f50f2600-78a9-4175-b0af-c5a18563e39d-audit-policies\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.055742 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.055828 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.055906 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.056005 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.056096 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.056176 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-user-template-error\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.056261 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-service-ca\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.056341 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-session\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.056422 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-router-certs\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.056508 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.139831 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.146511 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.157602 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-session\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.157647 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-router-certs\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.157675 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.157709 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-user-template-login\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.157758 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ct67\" (UniqueName: \"kubernetes.io/projected/f50f2600-78a9-4175-b0af-c5a18563e39d-kube-api-access-2ct67\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.158197 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f50f2600-78a9-4175-b0af-c5a18563e39d-audit-dir\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.158223 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f50f2600-78a9-4175-b0af-c5a18563e39d-audit-policies\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.158285 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f50f2600-78a9-4175-b0af-c5a18563e39d-audit-dir\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.158445 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.159027 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f50f2600-78a9-4175-b0af-c5a18563e39d-audit-policies\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.159145 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.159590 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.159648 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.159691 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.159712 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.160454 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-user-template-error\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.160588 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.160589 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-service-ca\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.161382 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-service-ca\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.163405 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-session\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.165209 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.165554 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-user-template-login\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.165778 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.166096 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-router-certs\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.166877 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.167186 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-user-template-error\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.175588 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f50f2600-78a9-4175-b0af-c5a18563e39d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.179517 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ct67\" (UniqueName: \"kubernetes.io/projected/f50f2600-78a9-4175-b0af-c5a18563e39d-kube-api-access-2ct67\") pod \"oauth-openshift-ccc74cc7-zj2l6\" (UID: \"f50f2600-78a9-4175-b0af-c5a18563e39d\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.189775 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.191561 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s" event={"ID":"ee6b0713-849b-4108-8fa0-53739ec95458","Type":"ContainerDied","Data":"93b365df74e3827364246978c264e3d204e2bd1996d2aefbc8d5def5272341e2"} Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.191626 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93b365df74e3827364246978c264e3d204e2bd1996d2aefbc8d5def5272341e2" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.195545 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.259668 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.307603 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.317385 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.375700 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.476416 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.540859 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.606473 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.633540 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.651210 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.665358 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.727821 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.738995 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-ccc74cc7-zj2l6"] Dec 01 17:30:16 crc kubenswrapper[4868]: W1201 17:30:16.749229 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf50f2600_78a9_4175_b0af_c5a18563e39d.slice/crio-205d68e6ad41a42c48d2a62ea4b3f7a821798e0315948df0214a7f6c22403184 WatchSource:0}: Error finding container 205d68e6ad41a42c48d2a62ea4b3f7a821798e0315948df0214a7f6c22403184: Status 404 returned error can't find the container with id 205d68e6ad41a42c48d2a62ea4b3f7a821798e0315948df0214a7f6c22403184 Dec 01 17:30:16 crc kubenswrapper[4868]: I1201 17:30:16.866441 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.054916 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.077220 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.124166 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.128512 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.197332 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" event={"ID":"f50f2600-78a9-4175-b0af-c5a18563e39d","Type":"ContainerStarted","Data":"0fb4cf32c12f3451aa82536d70a90872667c4b916afa5c5702cbffd01a641e34"} Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.197394 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" event={"ID":"f50f2600-78a9-4175-b0af-c5a18563e39d","Type":"ContainerStarted","Data":"205d68e6ad41a42c48d2a62ea4b3f7a821798e0315948df0214a7f6c22403184"} Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.197631 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.227716 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" podStartSLOduration=49.227699585 podStartE2EDuration="49.227699585s" podCreationTimestamp="2025-12-01 17:29:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:30:17.226741527 +0000 UTC m=+289.597851938" watchObservedRunningTime="2025-12-01 17:30:17.227699585 +0000 UTC m=+289.598809996" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.228927 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.235287 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.236686 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.295436 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.438290 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.483796 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.531720 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.606586 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.710482 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-ccc74cc7-zj2l6" Dec 01 17:30:17 crc kubenswrapper[4868]: I1201 17:30:17.864129 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 17:30:18 crc kubenswrapper[4868]: I1201 17:30:18.006179 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 17:30:18 crc kubenswrapper[4868]: I1201 17:30:18.032684 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 17:30:18 crc kubenswrapper[4868]: I1201 17:30:18.123262 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 17:30:18 crc kubenswrapper[4868]: I1201 17:30:18.324892 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 17:30:18 crc kubenswrapper[4868]: I1201 17:30:18.479028 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 17:30:18 crc kubenswrapper[4868]: I1201 17:30:18.489992 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 17:30:18 crc kubenswrapper[4868]: I1201 17:30:18.539811 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 17:30:18 crc kubenswrapper[4868]: I1201 17:30:18.617726 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 17:30:18 crc kubenswrapper[4868]: I1201 17:30:18.732167 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 17:30:18 crc kubenswrapper[4868]: I1201 17:30:18.757493 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 17:30:18 crc kubenswrapper[4868]: I1201 17:30:18.805088 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 17:30:18 crc kubenswrapper[4868]: I1201 17:30:18.877582 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 17:30:18 crc kubenswrapper[4868]: I1201 17:30:18.881979 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 17:30:19 crc kubenswrapper[4868]: I1201 17:30:19.279513 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 17:30:19 crc kubenswrapper[4868]: I1201 17:30:19.698834 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 17:30:19 crc kubenswrapper[4868]: I1201 17:30:19.959397 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 17:30:20 crc kubenswrapper[4868]: I1201 17:30:20.618693 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 17:30:20 crc kubenswrapper[4868]: I1201 17:30:20.801303 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.072251 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.072363 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.224802 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.225100 4868 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="179da301cbcf8aee82195bfae2f3fae1790e9cfcc06a0df19fefa0e4bc624ec9" exitCode=137 Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.225179 4868 scope.go:117] "RemoveContainer" containerID="179da301cbcf8aee82195bfae2f3fae1790e9cfcc06a0df19fefa0e4bc624ec9" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.225179 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.241634 4868 scope.go:117] "RemoveContainer" containerID="179da301cbcf8aee82195bfae2f3fae1790e9cfcc06a0df19fefa0e4bc624ec9" Dec 01 17:30:21 crc kubenswrapper[4868]: E1201 17:30:21.242153 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"179da301cbcf8aee82195bfae2f3fae1790e9cfcc06a0df19fefa0e4bc624ec9\": container with ID starting with 179da301cbcf8aee82195bfae2f3fae1790e9cfcc06a0df19fefa0e4bc624ec9 not found: ID does not exist" containerID="179da301cbcf8aee82195bfae2f3fae1790e9cfcc06a0df19fefa0e4bc624ec9" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.242202 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"179da301cbcf8aee82195bfae2f3fae1790e9cfcc06a0df19fefa0e4bc624ec9"} err="failed to get container status \"179da301cbcf8aee82195bfae2f3fae1790e9cfcc06a0df19fefa0e4bc624ec9\": rpc error: code = NotFound desc = could not find container \"179da301cbcf8aee82195bfae2f3fae1790e9cfcc06a0df19fefa0e4bc624ec9\": container with ID starting with 179da301cbcf8aee82195bfae2f3fae1790e9cfcc06a0df19fefa0e4bc624ec9 not found: ID does not exist" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.243632 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.243680 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.243799 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.243807 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.243857 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.243897 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.243926 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.243974 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.244066 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.244428 4868 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.244447 4868 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.244458 4868 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.244468 4868 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.250542 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:30:21 crc kubenswrapper[4868]: I1201 17:30:21.346551 4868 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:22 crc kubenswrapper[4868]: I1201 17:30:22.064825 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 17:30:22 crc kubenswrapper[4868]: I1201 17:30:22.181651 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.751119 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b89j6"] Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.751881 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b89j6" podUID="262076bc-ef10-4531-a6f9-f32148ee7a3e" containerName="registry-server" containerID="cri-o://00136348a5d6fe9d51d6e34c1e88a69b9216fb2ac25f3f9950a8b170dee2edb7" gracePeriod=30 Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.757252 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c9tpp"] Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.757691 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c9tpp" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" containerName="registry-server" containerID="cri-o://6c87d18bcef7c3d59b130c2225c87b56d162d5ac9e5ce3583898a2080579b5b5" gracePeriod=30 Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.767774 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xtcsm"] Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.768105 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xtcsm" podUID="b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" containerName="registry-server" containerID="cri-o://c2d7039d55006e07b57b82c23ad7262ead3c43e41c20b54382e22e41b5f8c8da" gracePeriod=30 Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.776028 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4hvgn"] Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.776259 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" podUID="22cc1879-5e5e-4a06-ba88-f92033382b90" containerName="marketplace-operator" containerID="cri-o://8e73eb20d2657fbf2df84c444f2df5c0d7a00fddee5dea5a79f5697a15825991" gracePeriod=30 Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.782051 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnrd7"] Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.782587 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fnrd7" podUID="de9baf7b-7da5-4cd3-908a-d3bc26f918f1" containerName="registry-server" containerID="cri-o://cc55d48ae37cb8986ddf62f4b7e17efc0b435d1be46faefeb4f8c57dcc5f5352" gracePeriod=30 Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.791009 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2f8vp"] Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.791314 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2f8vp" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" containerName="registry-server" containerID="cri-o://860fb316d47a18c7681dea9bb1438f501a3657b1fc3c566a1a43b02eb94837af" gracePeriod=30 Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.802820 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4nfp5"] Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.804527 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.812576 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pwdv6"] Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.812901 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pwdv6" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" containerName="registry-server" containerID="cri-o://b3b78e31c529d875946a4c9c212fdbbff5fdd7825c32083e081be87ff543b7bd" gracePeriod=30 Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.817333 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4nfp5"] Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.917137 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6qtn\" (UniqueName: \"kubernetes.io/projected/0c043a8c-1af1-4aa0-99f9-0f6ebe27238a-kube-api-access-t6qtn\") pod \"marketplace-operator-79b997595-4nfp5\" (UID: \"0c043a8c-1af1-4aa0-99f9-0f6ebe27238a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.917194 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0c043a8c-1af1-4aa0-99f9-0f6ebe27238a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4nfp5\" (UID: \"0c043a8c-1af1-4aa0-99f9-0f6ebe27238a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" Dec 01 17:30:25 crc kubenswrapper[4868]: I1201 17:30:25.917424 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0c043a8c-1af1-4aa0-99f9-0f6ebe27238a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4nfp5\" (UID: \"0c043a8c-1af1-4aa0-99f9-0f6ebe27238a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.019333 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6qtn\" (UniqueName: \"kubernetes.io/projected/0c043a8c-1af1-4aa0-99f9-0f6ebe27238a-kube-api-access-t6qtn\") pod \"marketplace-operator-79b997595-4nfp5\" (UID: \"0c043a8c-1af1-4aa0-99f9-0f6ebe27238a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.019383 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0c043a8c-1af1-4aa0-99f9-0f6ebe27238a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4nfp5\" (UID: \"0c043a8c-1af1-4aa0-99f9-0f6ebe27238a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.019435 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0c043a8c-1af1-4aa0-99f9-0f6ebe27238a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4nfp5\" (UID: \"0c043a8c-1af1-4aa0-99f9-0f6ebe27238a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.024384 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0c043a8c-1af1-4aa0-99f9-0f6ebe27238a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4nfp5\" (UID: \"0c043a8c-1af1-4aa0-99f9-0f6ebe27238a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.042706 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0c043a8c-1af1-4aa0-99f9-0f6ebe27238a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4nfp5\" (UID: \"0c043a8c-1af1-4aa0-99f9-0f6ebe27238a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.042764 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6qtn\" (UniqueName: \"kubernetes.io/projected/0c043a8c-1af1-4aa0-99f9-0f6ebe27238a-kube-api-access-t6qtn\") pod \"marketplace-operator-79b997595-4nfp5\" (UID: \"0c043a8c-1af1-4aa0-99f9-0f6ebe27238a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" Dec 01 17:30:26 crc kubenswrapper[4868]: E1201 17:30:26.213978 4868 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 00136348a5d6fe9d51d6e34c1e88a69b9216fb2ac25f3f9950a8b170dee2edb7 is running failed: container process not found" containerID="00136348a5d6fe9d51d6e34c1e88a69b9216fb2ac25f3f9950a8b170dee2edb7" cmd=["grpc_health_probe","-addr=:50051"] Dec 01 17:30:26 crc kubenswrapper[4868]: E1201 17:30:26.214815 4868 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 00136348a5d6fe9d51d6e34c1e88a69b9216fb2ac25f3f9950a8b170dee2edb7 is running failed: container process not found" containerID="00136348a5d6fe9d51d6e34c1e88a69b9216fb2ac25f3f9950a8b170dee2edb7" cmd=["grpc_health_probe","-addr=:50051"] Dec 01 17:30:26 crc kubenswrapper[4868]: E1201 17:30:26.215362 4868 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 00136348a5d6fe9d51d6e34c1e88a69b9216fb2ac25f3f9950a8b170dee2edb7 is running failed: container process not found" containerID="00136348a5d6fe9d51d6e34c1e88a69b9216fb2ac25f3f9950a8b170dee2edb7" cmd=["grpc_health_probe","-addr=:50051"] Dec 01 17:30:26 crc kubenswrapper[4868]: E1201 17:30:26.215404 4868 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 00136348a5d6fe9d51d6e34c1e88a69b9216fb2ac25f3f9950a8b170dee2edb7 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-b89j6" podUID="262076bc-ef10-4531-a6f9-f32148ee7a3e" containerName="registry-server" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.261475 4868 generic.go:334] "Generic (PLEG): container finished" podID="de9baf7b-7da5-4cd3-908a-d3bc26f918f1" containerID="cc55d48ae37cb8986ddf62f4b7e17efc0b435d1be46faefeb4f8c57dcc5f5352" exitCode=0 Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.261556 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnrd7" event={"ID":"de9baf7b-7da5-4cd3-908a-d3bc26f918f1","Type":"ContainerDied","Data":"cc55d48ae37cb8986ddf62f4b7e17efc0b435d1be46faefeb4f8c57dcc5f5352"} Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.269281 4868 generic.go:334] "Generic (PLEG): container finished" podID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" containerID="860fb316d47a18c7681dea9bb1438f501a3657b1fc3c566a1a43b02eb94837af" exitCode=0 Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.269365 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2f8vp" event={"ID":"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e","Type":"ContainerDied","Data":"860fb316d47a18c7681dea9bb1438f501a3657b1fc3c566a1a43b02eb94837af"} Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.271980 4868 generic.go:334] "Generic (PLEG): container finished" podID="262076bc-ef10-4531-a6f9-f32148ee7a3e" containerID="00136348a5d6fe9d51d6e34c1e88a69b9216fb2ac25f3f9950a8b170dee2edb7" exitCode=0 Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.272049 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b89j6" event={"ID":"262076bc-ef10-4531-a6f9-f32148ee7a3e","Type":"ContainerDied","Data":"00136348a5d6fe9d51d6e34c1e88a69b9216fb2ac25f3f9950a8b170dee2edb7"} Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.272085 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b89j6" event={"ID":"262076bc-ef10-4531-a6f9-f32148ee7a3e","Type":"ContainerDied","Data":"13f130cd42622905a02755a1d0729f1ea252cfeaf58d815d802363eab0b52908"} Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.272099 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13f130cd42622905a02755a1d0729f1ea252cfeaf58d815d802363eab0b52908" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.274328 4868 generic.go:334] "Generic (PLEG): container finished" podID="b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" containerID="c2d7039d55006e07b57b82c23ad7262ead3c43e41c20b54382e22e41b5f8c8da" exitCode=0 Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.274354 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xtcsm" event={"ID":"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06","Type":"ContainerDied","Data":"c2d7039d55006e07b57b82c23ad7262ead3c43e41c20b54382e22e41b5f8c8da"} Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.276819 4868 generic.go:334] "Generic (PLEG): container finished" podID="6332f65b-5f50-4e61-9f2d-b4333fda483b" containerID="b3b78e31c529d875946a4c9c212fdbbff5fdd7825c32083e081be87ff543b7bd" exitCode=0 Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.276869 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwdv6" event={"ID":"6332f65b-5f50-4e61-9f2d-b4333fda483b","Type":"ContainerDied","Data":"b3b78e31c529d875946a4c9c212fdbbff5fdd7825c32083e081be87ff543b7bd"} Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.278207 4868 generic.go:334] "Generic (PLEG): container finished" podID="22cc1879-5e5e-4a06-ba88-f92033382b90" containerID="8e73eb20d2657fbf2df84c444f2df5c0d7a00fddee5dea5a79f5697a15825991" exitCode=0 Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.278262 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" event={"ID":"22cc1879-5e5e-4a06-ba88-f92033382b90","Type":"ContainerDied","Data":"8e73eb20d2657fbf2df84c444f2df5c0d7a00fddee5dea5a79f5697a15825991"} Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.281296 4868 generic.go:334] "Generic (PLEG): container finished" podID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" containerID="6c87d18bcef7c3d59b130c2225c87b56d162d5ac9e5ce3583898a2080579b5b5" exitCode=0 Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.281320 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9tpp" event={"ID":"b2c7b9b4-a824-4639-9dfa-21778b900c9a","Type":"ContainerDied","Data":"6c87d18bcef7c3d59b130c2225c87b56d162d5ac9e5ce3583898a2080579b5b5"} Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.356364 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.363527 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.368461 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.376045 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.384229 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.390413 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.416295 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.445582 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.528961 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2c7b9b4-a824-4639-9dfa-21778b900c9a-utilities\") pod \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\" (UID: \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529024 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-catalog-content\") pod \"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e\" (UID: \"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529069 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7qrp\" (UniqueName: \"kubernetes.io/projected/262076bc-ef10-4531-a6f9-f32148ee7a3e-kube-api-access-n7qrp\") pod \"262076bc-ef10-4531-a6f9-f32148ee7a3e\" (UID: \"262076bc-ef10-4531-a6f9-f32148ee7a3e\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529105 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwg6w\" (UniqueName: \"kubernetes.io/projected/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-kube-api-access-jwg6w\") pod \"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e\" (UID: \"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529133 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-catalog-content\") pod \"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06\" (UID: \"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529175 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgls8\" (UniqueName: \"kubernetes.io/projected/6332f65b-5f50-4e61-9f2d-b4333fda483b-kube-api-access-pgls8\") pod \"6332f65b-5f50-4e61-9f2d-b4333fda483b\" (UID: \"6332f65b-5f50-4e61-9f2d-b4333fda483b\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529232 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nqj9\" (UniqueName: \"kubernetes.io/projected/22cc1879-5e5e-4a06-ba88-f92033382b90-kube-api-access-9nqj9\") pod \"22cc1879-5e5e-4a06-ba88-f92033382b90\" (UID: \"22cc1879-5e5e-4a06-ba88-f92033382b90\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529278 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22cc1879-5e5e-4a06-ba88-f92033382b90-marketplace-trusted-ca\") pod \"22cc1879-5e5e-4a06-ba88-f92033382b90\" (UID: \"22cc1879-5e5e-4a06-ba88-f92033382b90\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529313 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2c7b9b4-a824-4639-9dfa-21778b900c9a-catalog-content\") pod \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\" (UID: \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529340 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-utilities\") pod \"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e\" (UID: \"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529589 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-utilities\") pod \"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06\" (UID: \"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529630 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/22cc1879-5e5e-4a06-ba88-f92033382b90-marketplace-operator-metrics\") pod \"22cc1879-5e5e-4a06-ba88-f92033382b90\" (UID: \"22cc1879-5e5e-4a06-ba88-f92033382b90\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529652 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/262076bc-ef10-4531-a6f9-f32148ee7a3e-catalog-content\") pod \"262076bc-ef10-4531-a6f9-f32148ee7a3e\" (UID: \"262076bc-ef10-4531-a6f9-f32148ee7a3e\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529694 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6332f65b-5f50-4e61-9f2d-b4333fda483b-catalog-content\") pod \"6332f65b-5f50-4e61-9f2d-b4333fda483b\" (UID: \"6332f65b-5f50-4e61-9f2d-b4333fda483b\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529718 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6332f65b-5f50-4e61-9f2d-b4333fda483b-utilities\") pod \"6332f65b-5f50-4e61-9f2d-b4333fda483b\" (UID: \"6332f65b-5f50-4e61-9f2d-b4333fda483b\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529862 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvxwm\" (UniqueName: \"kubernetes.io/projected/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-kube-api-access-pvxwm\") pod \"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06\" (UID: \"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529894 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4b8mw\" (UniqueName: \"kubernetes.io/projected/b2c7b9b4-a824-4639-9dfa-21778b900c9a-kube-api-access-4b8mw\") pod \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\" (UID: \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.529920 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/262076bc-ef10-4531-a6f9-f32148ee7a3e-utilities\") pod \"262076bc-ef10-4531-a6f9-f32148ee7a3e\" (UID: \"262076bc-ef10-4531-a6f9-f32148ee7a3e\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.533252 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6332f65b-5f50-4e61-9f2d-b4333fda483b-utilities" (OuterVolumeSpecName: "utilities") pod "6332f65b-5f50-4e61-9f2d-b4333fda483b" (UID: "6332f65b-5f50-4e61-9f2d-b4333fda483b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.533291 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-utilities" (OuterVolumeSpecName: "utilities") pod "1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" (UID: "1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.533480 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-utilities" (OuterVolumeSpecName: "utilities") pod "b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" (UID: "b1529d6b-804b-4fb6-89a2-53cb5e2c5f06"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.538801 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22cc1879-5e5e-4a06-ba88-f92033382b90-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "22cc1879-5e5e-4a06-ba88-f92033382b90" (UID: "22cc1879-5e5e-4a06-ba88-f92033382b90"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.541500 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/262076bc-ef10-4531-a6f9-f32148ee7a3e-kube-api-access-n7qrp" (OuterVolumeSpecName: "kube-api-access-n7qrp") pod "262076bc-ef10-4531-a6f9-f32148ee7a3e" (UID: "262076bc-ef10-4531-a6f9-f32148ee7a3e"). InnerVolumeSpecName "kube-api-access-n7qrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.542136 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/262076bc-ef10-4531-a6f9-f32148ee7a3e-utilities" (OuterVolumeSpecName: "utilities") pod "262076bc-ef10-4531-a6f9-f32148ee7a3e" (UID: "262076bc-ef10-4531-a6f9-f32148ee7a3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.543560 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6332f65b-5f50-4e61-9f2d-b4333fda483b-kube-api-access-pgls8" (OuterVolumeSpecName: "kube-api-access-pgls8") pod "6332f65b-5f50-4e61-9f2d-b4333fda483b" (UID: "6332f65b-5f50-4e61-9f2d-b4333fda483b"). InnerVolumeSpecName "kube-api-access-pgls8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.543715 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2c7b9b4-a824-4639-9dfa-21778b900c9a-kube-api-access-4b8mw" (OuterVolumeSpecName: "kube-api-access-4b8mw") pod "b2c7b9b4-a824-4639-9dfa-21778b900c9a" (UID: "b2c7b9b4-a824-4639-9dfa-21778b900c9a"). InnerVolumeSpecName "kube-api-access-4b8mw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.544080 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2c7b9b4-a824-4639-9dfa-21778b900c9a-utilities" (OuterVolumeSpecName: "utilities") pod "b2c7b9b4-a824-4639-9dfa-21778b900c9a" (UID: "b2c7b9b4-a824-4639-9dfa-21778b900c9a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.545089 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22cc1879-5e5e-4a06-ba88-f92033382b90-kube-api-access-9nqj9" (OuterVolumeSpecName: "kube-api-access-9nqj9") pod "22cc1879-5e5e-4a06-ba88-f92033382b90" (UID: "22cc1879-5e5e-4a06-ba88-f92033382b90"). InnerVolumeSpecName "kube-api-access-9nqj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.555917 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-kube-api-access-jwg6w" (OuterVolumeSpecName: "kube-api-access-jwg6w") pod "1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" (UID: "1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e"). InnerVolumeSpecName "kube-api-access-jwg6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.562053 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22cc1879-5e5e-4a06-ba88-f92033382b90-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "22cc1879-5e5e-4a06-ba88-f92033382b90" (UID: "22cc1879-5e5e-4a06-ba88-f92033382b90"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.565294 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-kube-api-access-pvxwm" (OuterVolumeSpecName: "kube-api-access-pvxwm") pod "b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" (UID: "b1529d6b-804b-4fb6-89a2-53cb5e2c5f06"). InnerVolumeSpecName "kube-api-access-pvxwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.608371 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/262076bc-ef10-4531-a6f9-f32148ee7a3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "262076bc-ef10-4531-a6f9-f32148ee7a3e" (UID: "262076bc-ef10-4531-a6f9-f32148ee7a3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.632808 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c64j5\" (UniqueName: \"kubernetes.io/projected/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-kube-api-access-c64j5\") pod \"de9baf7b-7da5-4cd3-908a-d3bc26f918f1\" (UID: \"de9baf7b-7da5-4cd3-908a-d3bc26f918f1\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.632917 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2c7b9b4-a824-4639-9dfa-21778b900c9a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2c7b9b4-a824-4639-9dfa-21778b900c9a" (UID: "b2c7b9b4-a824-4639-9dfa-21778b900c9a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633117 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-catalog-content\") pod \"de9baf7b-7da5-4cd3-908a-d3bc26f918f1\" (UID: \"de9baf7b-7da5-4cd3-908a-d3bc26f918f1\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633280 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-utilities\") pod \"de9baf7b-7da5-4cd3-908a-d3bc26f918f1\" (UID: \"de9baf7b-7da5-4cd3-908a-d3bc26f918f1\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633379 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2c7b9b4-a824-4639-9dfa-21778b900c9a-catalog-content\") pod \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\" (UID: \"b2c7b9b4-a824-4639-9dfa-21778b900c9a\") " Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633743 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgls8\" (UniqueName: \"kubernetes.io/projected/6332f65b-5f50-4e61-9f2d-b4333fda483b-kube-api-access-pgls8\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633762 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nqj9\" (UniqueName: \"kubernetes.io/projected/22cc1879-5e5e-4a06-ba88-f92033382b90-kube-api-access-9nqj9\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633773 4868 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22cc1879-5e5e-4a06-ba88-f92033382b90-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633785 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633814 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633823 4868 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/22cc1879-5e5e-4a06-ba88-f92033382b90-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633832 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/262076bc-ef10-4531-a6f9-f32148ee7a3e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633841 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6332f65b-5f50-4e61-9f2d-b4333fda483b-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633852 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvxwm\" (UniqueName: \"kubernetes.io/projected/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-kube-api-access-pvxwm\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633861 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4b8mw\" (UniqueName: \"kubernetes.io/projected/b2c7b9b4-a824-4639-9dfa-21778b900c9a-kube-api-access-4b8mw\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633889 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/262076bc-ef10-4531-a6f9-f32148ee7a3e-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633900 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2c7b9b4-a824-4639-9dfa-21778b900c9a-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633909 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7qrp\" (UniqueName: \"kubernetes.io/projected/262076bc-ef10-4531-a6f9-f32148ee7a3e-kube-api-access-n7qrp\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.633918 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwg6w\" (UniqueName: \"kubernetes.io/projected/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-kube-api-access-jwg6w\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: W1201 17:30:26.634045 4868 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/b2c7b9b4-a824-4639-9dfa-21778b900c9a/volumes/kubernetes.io~empty-dir/catalog-content Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.634062 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2c7b9b4-a824-4639-9dfa-21778b900c9a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2c7b9b4-a824-4639-9dfa-21778b900c9a" (UID: "b2c7b9b4-a824-4639-9dfa-21778b900c9a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.635840 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" (UID: "b1529d6b-804b-4fb6-89a2-53cb5e2c5f06"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.637852 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-utilities" (OuterVolumeSpecName: "utilities") pod "de9baf7b-7da5-4cd3-908a-d3bc26f918f1" (UID: "de9baf7b-7da5-4cd3-908a-d3bc26f918f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.640130 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-kube-api-access-c64j5" (OuterVolumeSpecName: "kube-api-access-c64j5") pod "de9baf7b-7da5-4cd3-908a-d3bc26f918f1" (UID: "de9baf7b-7da5-4cd3-908a-d3bc26f918f1"). InnerVolumeSpecName "kube-api-access-c64j5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.654416 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de9baf7b-7da5-4cd3-908a-d3bc26f918f1" (UID: "de9baf7b-7da5-4cd3-908a-d3bc26f918f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.696339 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" (UID: "1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.703274 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6332f65b-5f50-4e61-9f2d-b4333fda483b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6332f65b-5f50-4e61-9f2d-b4333fda483b" (UID: "6332f65b-5f50-4e61-9f2d-b4333fda483b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.735535 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.735585 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.735604 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.735616 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2c7b9b4-a824-4639-9dfa-21778b900c9a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.735632 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c64j5\" (UniqueName: \"kubernetes.io/projected/de9baf7b-7da5-4cd3-908a-d3bc26f918f1-kube-api-access-c64j5\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.735647 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6332f65b-5f50-4e61-9f2d-b4333fda483b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.735659 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:26 crc kubenswrapper[4868]: I1201 17:30:26.835149 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4nfp5"] Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.290639 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9tpp" event={"ID":"b2c7b9b4-a824-4639-9dfa-21778b900c9a","Type":"ContainerDied","Data":"553e5dadc5b0e3d510394c44df2681b51d2f9ffc58f071004dd3aedea85f5c66"} Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.291137 4868 scope.go:117] "RemoveContainer" containerID="6c87d18bcef7c3d59b130c2225c87b56d162d5ac9e5ce3583898a2080579b5b5" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.290658 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c9tpp" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.293026 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnrd7" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.293139 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnrd7" event={"ID":"de9baf7b-7da5-4cd3-908a-d3bc26f918f1","Type":"ContainerDied","Data":"0b1ffba19d22107b3ab1498e7c1358b95d9dd1fd97f8609949d873f9f3a2ff05"} Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.296166 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2f8vp" event={"ID":"1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e","Type":"ContainerDied","Data":"a8a9cc3beaef944fd48ce80d52139b000a3ddab29129ae8fb45f9bfa1de8f6ae"} Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.296361 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2f8vp" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.302879 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwdv6" event={"ID":"6332f65b-5f50-4e61-9f2d-b4333fda483b","Type":"ContainerDied","Data":"f2ed393f6a86b1e6a801d8a17df8357df3bd31c9277716bd7778bb07f4a5bdc3"} Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.303037 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwdv6" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.306976 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" event={"ID":"0c043a8c-1af1-4aa0-99f9-0f6ebe27238a","Type":"ContainerStarted","Data":"d1391a60b4ed985657c05ce04447cce62b8925b0c4c2ea85d08c332e8fda8622"} Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.307026 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" event={"ID":"0c043a8c-1af1-4aa0-99f9-0f6ebe27238a","Type":"ContainerStarted","Data":"1fd96d52619ad1521870ff0b25386b79b03160188cbacede30b91535f8415c03"} Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.307609 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.311064 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" event={"ID":"22cc1879-5e5e-4a06-ba88-f92033382b90","Type":"ContainerDied","Data":"bbc91beeabbd54e17c194ad29b3c73dde0a7a4b3f01415f83582bb754891b950"} Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.311141 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4hvgn" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.313640 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.319596 4868 scope.go:117] "RemoveContainer" containerID="e8919dbc7948177a59fe00b4211e5c84e9c7b11b4c8fa41f0868b28933392fe4" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.320066 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xtcsm" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.320042 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xtcsm" event={"ID":"b1529d6b-804b-4fb6-89a2-53cb5e2c5f06","Type":"ContainerDied","Data":"a4dc3c532e7ba228c311a079eb860960773fe3fddb0b5e782ea0af8da69a4c7b"} Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.320180 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b89j6" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.328118 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4nfp5" podStartSLOduration=2.328093213 podStartE2EDuration="2.328093213s" podCreationTimestamp="2025-12-01 17:30:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:30:27.326099754 +0000 UTC m=+299.697210205" watchObservedRunningTime="2025-12-01 17:30:27.328093213 +0000 UTC m=+299.699203634" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.382211 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnrd7"] Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.395861 4868 scope.go:117] "RemoveContainer" containerID="cb9d980c782a5777e72b3ce43d7000fe2a846d22a628e55ea15ad130389c8cd6" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.397239 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnrd7"] Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.415427 4868 scope.go:117] "RemoveContainer" containerID="cc55d48ae37cb8986ddf62f4b7e17efc0b435d1be46faefeb4f8c57dcc5f5352" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.418285 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c9tpp"] Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.422971 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c9tpp"] Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.435254 4868 scope.go:117] "RemoveContainer" containerID="3207bdbc0f84b42cdc537b8bffcc1001bcca64e782d5115aecbfca764eb604a3" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.436635 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pwdv6"] Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.443194 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pwdv6"] Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.449078 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4hvgn"] Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.450333 4868 scope.go:117] "RemoveContainer" containerID="87b15060fb50c05e6ee2ee304b4c43aaf0c6591ab275f4561e0afb8d32be4b8c" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.456350 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4hvgn"] Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.462652 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xtcsm"] Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.467152 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xtcsm"] Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.474060 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2f8vp"] Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.477031 4868 scope.go:117] "RemoveContainer" containerID="860fb316d47a18c7681dea9bb1438f501a3657b1fc3c566a1a43b02eb94837af" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.479122 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2f8vp"] Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.502730 4868 scope.go:117] "RemoveContainer" containerID="b46685901acab86a775f40a7851a7fa0412e42b9eb1d66c2b21931cb4b6ab4b5" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.508979 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b89j6"] Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.518440 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b89j6"] Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.519977 4868 scope.go:117] "RemoveContainer" containerID="d6d11c99b8802ef5c4d80b716e301094f0a7c3fc00b0ca3a66f411a679c21f07" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.537020 4868 scope.go:117] "RemoveContainer" containerID="b3b78e31c529d875946a4c9c212fdbbff5fdd7825c32083e081be87ff543b7bd" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.553203 4868 scope.go:117] "RemoveContainer" containerID="747f002d75adf979a20407ec8852e9555659ca1af2c7068335be9a21848d29ca" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.571000 4868 scope.go:117] "RemoveContainer" containerID="cf302c29f1b56ee7be86d0e4fb9989908674e311b4acb83ec0035d1d6b2bf9aa" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.586873 4868 scope.go:117] "RemoveContainer" containerID="8e73eb20d2657fbf2df84c444f2df5c0d7a00fddee5dea5a79f5697a15825991" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.602073 4868 scope.go:117] "RemoveContainer" containerID="c2d7039d55006e07b57b82c23ad7262ead3c43e41c20b54382e22e41b5f8c8da" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.619576 4868 scope.go:117] "RemoveContainer" containerID="2e84e511a4b3d6e3bc4fde7cc484e08f0239dad6a9374c13c8dd6e73a2bdb1c8" Dec 01 17:30:27 crc kubenswrapper[4868]: I1201 17:30:27.635337 4868 scope.go:117] "RemoveContainer" containerID="afe3497f1a9177edd7026915922c659b3dc467b09e9a055b2765bcdd7fdae125" Dec 01 17:30:28 crc kubenswrapper[4868]: I1201 17:30:28.180889 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" path="/var/lib/kubelet/pods/1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e/volumes" Dec 01 17:30:28 crc kubenswrapper[4868]: I1201 17:30:28.181933 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22cc1879-5e5e-4a06-ba88-f92033382b90" path="/var/lib/kubelet/pods/22cc1879-5e5e-4a06-ba88-f92033382b90/volumes" Dec 01 17:30:28 crc kubenswrapper[4868]: I1201 17:30:28.183516 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="262076bc-ef10-4531-a6f9-f32148ee7a3e" path="/var/lib/kubelet/pods/262076bc-ef10-4531-a6f9-f32148ee7a3e/volumes" Dec 01 17:30:28 crc kubenswrapper[4868]: I1201 17:30:28.184529 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" path="/var/lib/kubelet/pods/6332f65b-5f50-4e61-9f2d-b4333fda483b/volumes" Dec 01 17:30:28 crc kubenswrapper[4868]: I1201 17:30:28.186303 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" path="/var/lib/kubelet/pods/b1529d6b-804b-4fb6-89a2-53cb5e2c5f06/volumes" Dec 01 17:30:28 crc kubenswrapper[4868]: I1201 17:30:28.187331 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" path="/var/lib/kubelet/pods/b2c7b9b4-a824-4639-9dfa-21778b900c9a/volumes" Dec 01 17:30:28 crc kubenswrapper[4868]: I1201 17:30:28.188372 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de9baf7b-7da5-4cd3-908a-d3bc26f918f1" path="/var/lib/kubelet/pods/de9baf7b-7da5-4cd3-908a-d3bc26f918f1/volumes" Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.459184 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-s4hdj"] Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.459935 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" podUID="b124c538-2e03-45ea-9666-b52b1fc90bdb" containerName="controller-manager" containerID="cri-o://902a78ae3ce382dd36022c1cbe55a8af2c736995e0d92b8f0fe25eec50dd7da5" gracePeriod=30 Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.575390 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6"] Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.575740 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" podUID="a743581d-083b-4400-a1f3-a73f1cfece31" containerName="route-controller-manager" containerID="cri-o://3c5284e965f3be4e86608e4bd1301c2443e174e3ecc7737c958454d7d9c9675c" gracePeriod=30 Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.868101 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.950800 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tjdp\" (UniqueName: \"kubernetes.io/projected/b124c538-2e03-45ea-9666-b52b1fc90bdb-kube-api-access-2tjdp\") pod \"b124c538-2e03-45ea-9666-b52b1fc90bdb\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.950888 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-client-ca\") pod \"b124c538-2e03-45ea-9666-b52b1fc90bdb\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.950967 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-proxy-ca-bundles\") pod \"b124c538-2e03-45ea-9666-b52b1fc90bdb\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.951040 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-config\") pod \"b124c538-2e03-45ea-9666-b52b1fc90bdb\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.951072 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b124c538-2e03-45ea-9666-b52b1fc90bdb-serving-cert\") pod \"b124c538-2e03-45ea-9666-b52b1fc90bdb\" (UID: \"b124c538-2e03-45ea-9666-b52b1fc90bdb\") " Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.953021 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-client-ca" (OuterVolumeSpecName: "client-ca") pod "b124c538-2e03-45ea-9666-b52b1fc90bdb" (UID: "b124c538-2e03-45ea-9666-b52b1fc90bdb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.953367 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b124c538-2e03-45ea-9666-b52b1fc90bdb" (UID: "b124c538-2e03-45ea-9666-b52b1fc90bdb"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.953827 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-config" (OuterVolumeSpecName: "config") pod "b124c538-2e03-45ea-9666-b52b1fc90bdb" (UID: "b124c538-2e03-45ea-9666-b52b1fc90bdb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.959766 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b124c538-2e03-45ea-9666-b52b1fc90bdb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b124c538-2e03-45ea-9666-b52b1fc90bdb" (UID: "b124c538-2e03-45ea-9666-b52b1fc90bdb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.960410 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b124c538-2e03-45ea-9666-b52b1fc90bdb-kube-api-access-2tjdp" (OuterVolumeSpecName: "kube-api-access-2tjdp") pod "b124c538-2e03-45ea-9666-b52b1fc90bdb" (UID: "b124c538-2e03-45ea-9666-b52b1fc90bdb"). InnerVolumeSpecName "kube-api-access-2tjdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:30:51 crc kubenswrapper[4868]: I1201 17:30:51.998301 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.052440 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a743581d-083b-4400-a1f3-a73f1cfece31-client-ca\") pod \"a743581d-083b-4400-a1f3-a73f1cfece31\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.052513 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a743581d-083b-4400-a1f3-a73f1cfece31-config\") pod \"a743581d-083b-4400-a1f3-a73f1cfece31\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.052561 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a743581d-083b-4400-a1f3-a73f1cfece31-serving-cert\") pod \"a743581d-083b-4400-a1f3-a73f1cfece31\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.052718 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcv9g\" (UniqueName: \"kubernetes.io/projected/a743581d-083b-4400-a1f3-a73f1cfece31-kube-api-access-jcv9g\") pod \"a743581d-083b-4400-a1f3-a73f1cfece31\" (UID: \"a743581d-083b-4400-a1f3-a73f1cfece31\") " Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.052958 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.052975 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b124c538-2e03-45ea-9666-b52b1fc90bdb-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.052985 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tjdp\" (UniqueName: \"kubernetes.io/projected/b124c538-2e03-45ea-9666-b52b1fc90bdb-kube-api-access-2tjdp\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.052997 4868 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.053011 4868 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b124c538-2e03-45ea-9666-b52b1fc90bdb-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.053504 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a743581d-083b-4400-a1f3-a73f1cfece31-client-ca" (OuterVolumeSpecName: "client-ca") pod "a743581d-083b-4400-a1f3-a73f1cfece31" (UID: "a743581d-083b-4400-a1f3-a73f1cfece31"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.053642 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a743581d-083b-4400-a1f3-a73f1cfece31-config" (OuterVolumeSpecName: "config") pod "a743581d-083b-4400-a1f3-a73f1cfece31" (UID: "a743581d-083b-4400-a1f3-a73f1cfece31"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.056529 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a743581d-083b-4400-a1f3-a73f1cfece31-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a743581d-083b-4400-a1f3-a73f1cfece31" (UID: "a743581d-083b-4400-a1f3-a73f1cfece31"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.056849 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a743581d-083b-4400-a1f3-a73f1cfece31-kube-api-access-jcv9g" (OuterVolumeSpecName: "kube-api-access-jcv9g") pod "a743581d-083b-4400-a1f3-a73f1cfece31" (UID: "a743581d-083b-4400-a1f3-a73f1cfece31"). InnerVolumeSpecName "kube-api-access-jcv9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.154538 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcv9g\" (UniqueName: \"kubernetes.io/projected/a743581d-083b-4400-a1f3-a73f1cfece31-kube-api-access-jcv9g\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.154594 4868 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a743581d-083b-4400-a1f3-a73f1cfece31-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.154605 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a743581d-083b-4400-a1f3-a73f1cfece31-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.154615 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a743581d-083b-4400-a1f3-a73f1cfece31-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.490596 4868 generic.go:334] "Generic (PLEG): container finished" podID="a743581d-083b-4400-a1f3-a73f1cfece31" containerID="3c5284e965f3be4e86608e4bd1301c2443e174e3ecc7737c958454d7d9c9675c" exitCode=0 Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.490703 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" event={"ID":"a743581d-083b-4400-a1f3-a73f1cfece31","Type":"ContainerDied","Data":"3c5284e965f3be4e86608e4bd1301c2443e174e3ecc7737c958454d7d9c9675c"} Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.490754 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" event={"ID":"a743581d-083b-4400-a1f3-a73f1cfece31","Type":"ContainerDied","Data":"8d44d29517630feb4ed140233c1e936f90449d7f512e0c9d9831169ec1a9a8ad"} Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.490786 4868 scope.go:117] "RemoveContainer" containerID="3c5284e965f3be4e86608e4bd1301c2443e174e3ecc7737c958454d7d9c9675c" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.491008 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.496145 4868 generic.go:334] "Generic (PLEG): container finished" podID="b124c538-2e03-45ea-9666-b52b1fc90bdb" containerID="902a78ae3ce382dd36022c1cbe55a8af2c736995e0d92b8f0fe25eec50dd7da5" exitCode=0 Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.496221 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" event={"ID":"b124c538-2e03-45ea-9666-b52b1fc90bdb","Type":"ContainerDied","Data":"902a78ae3ce382dd36022c1cbe55a8af2c736995e0d92b8f0fe25eec50dd7da5"} Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.496264 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.496289 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-s4hdj" event={"ID":"b124c538-2e03-45ea-9666-b52b1fc90bdb","Type":"ContainerDied","Data":"21908cffef69ec16cb38ee5e8c1c45e04afbad581d2998ac1906b68df264afce"} Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.521705 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6"] Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.522590 4868 scope.go:117] "RemoveContainer" containerID="3c5284e965f3be4e86608e4bd1301c2443e174e3ecc7737c958454d7d9c9675c" Dec 01 17:30:52 crc kubenswrapper[4868]: E1201 17:30:52.525579 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c5284e965f3be4e86608e4bd1301c2443e174e3ecc7737c958454d7d9c9675c\": container with ID starting with 3c5284e965f3be4e86608e4bd1301c2443e174e3ecc7737c958454d7d9c9675c not found: ID does not exist" containerID="3c5284e965f3be4e86608e4bd1301c2443e174e3ecc7737c958454d7d9c9675c" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.525640 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c5284e965f3be4e86608e4bd1301c2443e174e3ecc7737c958454d7d9c9675c"} err="failed to get container status \"3c5284e965f3be4e86608e4bd1301c2443e174e3ecc7737c958454d7d9c9675c\": rpc error: code = NotFound desc = could not find container \"3c5284e965f3be4e86608e4bd1301c2443e174e3ecc7737c958454d7d9c9675c\": container with ID starting with 3c5284e965f3be4e86608e4bd1301c2443e174e3ecc7737c958454d7d9c9675c not found: ID does not exist" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.525678 4868 scope.go:117] "RemoveContainer" containerID="902a78ae3ce382dd36022c1cbe55a8af2c736995e0d92b8f0fe25eec50dd7da5" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.527008 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qc4x6"] Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.538244 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-s4hdj"] Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.542542 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-s4hdj"] Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.553782 4868 scope.go:117] "RemoveContainer" containerID="902a78ae3ce382dd36022c1cbe55a8af2c736995e0d92b8f0fe25eec50dd7da5" Dec 01 17:30:52 crc kubenswrapper[4868]: E1201 17:30:52.554363 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"902a78ae3ce382dd36022c1cbe55a8af2c736995e0d92b8f0fe25eec50dd7da5\": container with ID starting with 902a78ae3ce382dd36022c1cbe55a8af2c736995e0d92b8f0fe25eec50dd7da5 not found: ID does not exist" containerID="902a78ae3ce382dd36022c1cbe55a8af2c736995e0d92b8f0fe25eec50dd7da5" Dec 01 17:30:52 crc kubenswrapper[4868]: I1201 17:30:52.554406 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"902a78ae3ce382dd36022c1cbe55a8af2c736995e0d92b8f0fe25eec50dd7da5"} err="failed to get container status \"902a78ae3ce382dd36022c1cbe55a8af2c736995e0d92b8f0fe25eec50dd7da5\": rpc error: code = NotFound desc = could not find container \"902a78ae3ce382dd36022c1cbe55a8af2c736995e0d92b8f0fe25eec50dd7da5\": container with ID starting with 902a78ae3ce382dd36022c1cbe55a8af2c736995e0d92b8f0fe25eec50dd7da5 not found: ID does not exist" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.017683 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-599f95c79c-h76w9"] Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018114 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a743581d-083b-4400-a1f3-a73f1cfece31" containerName="route-controller-manager" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018145 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="a743581d-083b-4400-a1f3-a73f1cfece31" containerName="route-controller-manager" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018176 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9baf7b-7da5-4cd3-908a-d3bc26f918f1" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018189 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9baf7b-7da5-4cd3-908a-d3bc26f918f1" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018210 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" containerName="extract-utilities" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018223 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" containerName="extract-utilities" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018239 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="262076bc-ef10-4531-a6f9-f32148ee7a3e" containerName="extract-content" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018251 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="262076bc-ef10-4531-a6f9-f32148ee7a3e" containerName="extract-content" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018268 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" containerName="extract-content" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018281 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" containerName="extract-content" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018297 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9baf7b-7da5-4cd3-908a-d3bc26f918f1" containerName="extract-content" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018308 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9baf7b-7da5-4cd3-908a-d3bc26f918f1" containerName="extract-content" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018327 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9baf7b-7da5-4cd3-908a-d3bc26f918f1" containerName="extract-utilities" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018340 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9baf7b-7da5-4cd3-908a-d3bc26f918f1" containerName="extract-utilities" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018354 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018365 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018385 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018397 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018418 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" containerName="extract-content" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018434 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" containerName="extract-content" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018455 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018472 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018497 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" containerName="extract-utilities" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018514 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" containerName="extract-utilities" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018530 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" containerName="extract-content" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018543 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" containerName="extract-content" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018561 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018573 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018602 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b124c538-2e03-45ea-9666-b52b1fc90bdb" containerName="controller-manager" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018615 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b124c538-2e03-45ea-9666-b52b1fc90bdb" containerName="controller-manager" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018635 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" containerName="extract-utilities" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018648 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" containerName="extract-utilities" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018668 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22cc1879-5e5e-4a06-ba88-f92033382b90" containerName="marketplace-operator" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018681 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="22cc1879-5e5e-4a06-ba88-f92033382b90" containerName="marketplace-operator" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018699 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" containerName="extract-content" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018711 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" containerName="extract-content" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018731 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" containerName="extract-utilities" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018743 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" containerName="extract-utilities" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018762 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="262076bc-ef10-4531-a6f9-f32148ee7a3e" containerName="extract-utilities" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018774 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="262076bc-ef10-4531-a6f9-f32148ee7a3e" containerName="extract-utilities" Dec 01 17:30:53 crc kubenswrapper[4868]: E1201 17:30:53.018793 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="262076bc-ef10-4531-a6f9-f32148ee7a3e" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.018805 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="262076bc-ef10-4531-a6f9-f32148ee7a3e" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.019019 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dbeabb7-9a90-4b2c-9e1b-3d8ac947874e" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.019050 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="6332f65b-5f50-4e61-9f2d-b4333fda483b" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.019070 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1529d6b-804b-4fb6-89a2-53cb5e2c5f06" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.019100 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="b124c538-2e03-45ea-9666-b52b1fc90bdb" containerName="controller-manager" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.019122 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="22cc1879-5e5e-4a06-ba88-f92033382b90" containerName="marketplace-operator" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.019143 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="262076bc-ef10-4531-a6f9-f32148ee7a3e" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.019162 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="a743581d-083b-4400-a1f3-a73f1cfece31" containerName="route-controller-manager" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.019181 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2c7b9b4-a824-4639-9dfa-21778b900c9a" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.019201 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9baf7b-7da5-4cd3-908a-d3bc26f918f1" containerName="registry-server" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.020050 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.029414 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.029637 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.030180 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.030420 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.030643 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.030758 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.032462 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk"] Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.035256 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.039356 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.039433 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.039631 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.040042 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.040192 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.041452 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.041499 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.050487 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-599f95c79c-h76w9"] Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.057374 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk"] Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.067072 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26db4f43-c152-44a3-be40-c00407d62fb0-serving-cert\") pod \"controller-manager-599f95c79c-h76w9\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.067221 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c7b3490-9561-4058-865a-7d53ca5112c5-serving-cert\") pod \"route-controller-manager-84bcd68b97-xz5jk\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.067257 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4kr7\" (UniqueName: \"kubernetes.io/projected/6c7b3490-9561-4058-865a-7d53ca5112c5-kube-api-access-z4kr7\") pod \"route-controller-manager-84bcd68b97-xz5jk\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.067286 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-client-ca\") pod \"controller-manager-599f95c79c-h76w9\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.067368 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-proxy-ca-bundles\") pod \"controller-manager-599f95c79c-h76w9\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.067399 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-config\") pod \"controller-manager-599f95c79c-h76w9\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.067429 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqcsg\" (UniqueName: \"kubernetes.io/projected/26db4f43-c152-44a3-be40-c00407d62fb0-kube-api-access-wqcsg\") pod \"controller-manager-599f95c79c-h76w9\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.067526 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c7b3490-9561-4058-865a-7d53ca5112c5-client-ca\") pod \"route-controller-manager-84bcd68b97-xz5jk\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.067589 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c7b3490-9561-4058-865a-7d53ca5112c5-config\") pod \"route-controller-manager-84bcd68b97-xz5jk\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.169419 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26db4f43-c152-44a3-be40-c00407d62fb0-serving-cert\") pod \"controller-manager-599f95c79c-h76w9\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.169978 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c7b3490-9561-4058-865a-7d53ca5112c5-serving-cert\") pod \"route-controller-manager-84bcd68b97-xz5jk\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.170041 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4kr7\" (UniqueName: \"kubernetes.io/projected/6c7b3490-9561-4058-865a-7d53ca5112c5-kube-api-access-z4kr7\") pod \"route-controller-manager-84bcd68b97-xz5jk\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.170099 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-client-ca\") pod \"controller-manager-599f95c79c-h76w9\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.170159 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-proxy-ca-bundles\") pod \"controller-manager-599f95c79c-h76w9\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.170207 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-config\") pod \"controller-manager-599f95c79c-h76w9\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.170259 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqcsg\" (UniqueName: \"kubernetes.io/projected/26db4f43-c152-44a3-be40-c00407d62fb0-kube-api-access-wqcsg\") pod \"controller-manager-599f95c79c-h76w9\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.170318 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c7b3490-9561-4058-865a-7d53ca5112c5-client-ca\") pod \"route-controller-manager-84bcd68b97-xz5jk\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.170370 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c7b3490-9561-4058-865a-7d53ca5112c5-config\") pod \"route-controller-manager-84bcd68b97-xz5jk\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.171490 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-client-ca\") pod \"controller-manager-599f95c79c-h76w9\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.172166 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c7b3490-9561-4058-865a-7d53ca5112c5-client-ca\") pod \"route-controller-manager-84bcd68b97-xz5jk\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.173306 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-proxy-ca-bundles\") pod \"controller-manager-599f95c79c-h76w9\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.173315 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c7b3490-9561-4058-865a-7d53ca5112c5-config\") pod \"route-controller-manager-84bcd68b97-xz5jk\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.173324 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-config\") pod \"controller-manager-599f95c79c-h76w9\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.176575 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26db4f43-c152-44a3-be40-c00407d62fb0-serving-cert\") pod \"controller-manager-599f95c79c-h76w9\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.178730 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c7b3490-9561-4058-865a-7d53ca5112c5-serving-cert\") pod \"route-controller-manager-84bcd68b97-xz5jk\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.190376 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4kr7\" (UniqueName: \"kubernetes.io/projected/6c7b3490-9561-4058-865a-7d53ca5112c5-kube-api-access-z4kr7\") pod \"route-controller-manager-84bcd68b97-xz5jk\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.195409 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqcsg\" (UniqueName: \"kubernetes.io/projected/26db4f43-c152-44a3-be40-c00407d62fb0-kube-api-access-wqcsg\") pod \"controller-manager-599f95c79c-h76w9\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.361795 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.373467 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.604185 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-599f95c79c-h76w9"] Dec 01 17:30:53 crc kubenswrapper[4868]: I1201 17:30:53.631848 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk"] Dec 01 17:30:54 crc kubenswrapper[4868]: I1201 17:30:54.179829 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a743581d-083b-4400-a1f3-a73f1cfece31" path="/var/lib/kubelet/pods/a743581d-083b-4400-a1f3-a73f1cfece31/volumes" Dec 01 17:30:54 crc kubenswrapper[4868]: I1201 17:30:54.180897 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b124c538-2e03-45ea-9666-b52b1fc90bdb" path="/var/lib/kubelet/pods/b124c538-2e03-45ea-9666-b52b1fc90bdb/volumes" Dec 01 17:30:54 crc kubenswrapper[4868]: I1201 17:30:54.515441 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" event={"ID":"26db4f43-c152-44a3-be40-c00407d62fb0","Type":"ContainerStarted","Data":"0014d7a143cbbc6afe75079e23a3ef9427b4d11f21253762f4d29aef535a61ab"} Dec 01 17:30:54 crc kubenswrapper[4868]: I1201 17:30:54.515982 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" event={"ID":"26db4f43-c152-44a3-be40-c00407d62fb0","Type":"ContainerStarted","Data":"a3d8b15adb2848467cfe46231c22e740519b0196d8b5b69344bbe68d5ed937ad"} Dec 01 17:30:54 crc kubenswrapper[4868]: I1201 17:30:54.516771 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:54 crc kubenswrapper[4868]: I1201 17:30:54.519228 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" event={"ID":"6c7b3490-9561-4058-865a-7d53ca5112c5","Type":"ContainerStarted","Data":"a797f4686fe9ef2de0f50ae7a0c6dc7222a7e6d08c288eee28e269c6459fc465"} Dec 01 17:30:54 crc kubenswrapper[4868]: I1201 17:30:54.519282 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" event={"ID":"6c7b3490-9561-4058-865a-7d53ca5112c5","Type":"ContainerStarted","Data":"203396121c96a6e22011e78b33f3a7475687e2a2e785fcff0b375cf013049da5"} Dec 01 17:30:54 crc kubenswrapper[4868]: I1201 17:30:54.519457 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:54 crc kubenswrapper[4868]: I1201 17:30:54.522156 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:30:54 crc kubenswrapper[4868]: I1201 17:30:54.525117 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:30:54 crc kubenswrapper[4868]: I1201 17:30:54.554002 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" podStartSLOduration=3.5539722620000003 podStartE2EDuration="3.553972262s" podCreationTimestamp="2025-12-01 17:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:30:54.549684961 +0000 UTC m=+326.920795382" watchObservedRunningTime="2025-12-01 17:30:54.553972262 +0000 UTC m=+326.925082683" Dec 01 17:30:54 crc kubenswrapper[4868]: I1201 17:30:54.655204 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" podStartSLOduration=3.655120234 podStartE2EDuration="3.655120234s" podCreationTimestamp="2025-12-01 17:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:30:54.652392157 +0000 UTC m=+327.023502558" watchObservedRunningTime="2025-12-01 17:30:54.655120234 +0000 UTC m=+327.026230645" Dec 01 17:31:11 crc kubenswrapper[4868]: I1201 17:31:11.462682 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk"] Dec 01 17:31:11 crc kubenswrapper[4868]: I1201 17:31:11.463600 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" podUID="6c7b3490-9561-4058-865a-7d53ca5112c5" containerName="route-controller-manager" containerID="cri-o://a797f4686fe9ef2de0f50ae7a0c6dc7222a7e6d08c288eee28e269c6459fc465" gracePeriod=30 Dec 01 17:31:11 crc kubenswrapper[4868]: I1201 17:31:11.626336 4868 generic.go:334] "Generic (PLEG): container finished" podID="6c7b3490-9561-4058-865a-7d53ca5112c5" containerID="a797f4686fe9ef2de0f50ae7a0c6dc7222a7e6d08c288eee28e269c6459fc465" exitCode=0 Dec 01 17:31:11 crc kubenswrapper[4868]: I1201 17:31:11.626391 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" event={"ID":"6c7b3490-9561-4058-865a-7d53ca5112c5","Type":"ContainerDied","Data":"a797f4686fe9ef2de0f50ae7a0c6dc7222a7e6d08c288eee28e269c6459fc465"} Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.494807 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.529450 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w"] Dec 01 17:31:12 crc kubenswrapper[4868]: E1201 17:31:12.529755 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c7b3490-9561-4058-865a-7d53ca5112c5" containerName="route-controller-manager" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.529772 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c7b3490-9561-4058-865a-7d53ca5112c5" containerName="route-controller-manager" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.529905 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c7b3490-9561-4058-865a-7d53ca5112c5" containerName="route-controller-manager" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.530418 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.544561 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w"] Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.587175 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4kr7\" (UniqueName: \"kubernetes.io/projected/6c7b3490-9561-4058-865a-7d53ca5112c5-kube-api-access-z4kr7\") pod \"6c7b3490-9561-4058-865a-7d53ca5112c5\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.587332 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c7b3490-9561-4058-865a-7d53ca5112c5-serving-cert\") pod \"6c7b3490-9561-4058-865a-7d53ca5112c5\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.587464 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c7b3490-9561-4058-865a-7d53ca5112c5-client-ca\") pod \"6c7b3490-9561-4058-865a-7d53ca5112c5\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.587510 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c7b3490-9561-4058-865a-7d53ca5112c5-config\") pod \"6c7b3490-9561-4058-865a-7d53ca5112c5\" (UID: \"6c7b3490-9561-4058-865a-7d53ca5112c5\") " Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.589607 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c7b3490-9561-4058-865a-7d53ca5112c5-client-ca" (OuterVolumeSpecName: "client-ca") pod "6c7b3490-9561-4058-865a-7d53ca5112c5" (UID: "6c7b3490-9561-4058-865a-7d53ca5112c5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.589714 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c7b3490-9561-4058-865a-7d53ca5112c5-config" (OuterVolumeSpecName: "config") pod "6c7b3490-9561-4058-865a-7d53ca5112c5" (UID: "6c7b3490-9561-4058-865a-7d53ca5112c5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.597128 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c7b3490-9561-4058-865a-7d53ca5112c5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6c7b3490-9561-4058-865a-7d53ca5112c5" (UID: "6c7b3490-9561-4058-865a-7d53ca5112c5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.599329 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c7b3490-9561-4058-865a-7d53ca5112c5-kube-api-access-z4kr7" (OuterVolumeSpecName: "kube-api-access-z4kr7") pod "6c7b3490-9561-4058-865a-7d53ca5112c5" (UID: "6c7b3490-9561-4058-865a-7d53ca5112c5"). InnerVolumeSpecName "kube-api-access-z4kr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.637431 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" event={"ID":"6c7b3490-9561-4058-865a-7d53ca5112c5","Type":"ContainerDied","Data":"203396121c96a6e22011e78b33f3a7475687e2a2e785fcff0b375cf013049da5"} Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.637502 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.637510 4868 scope.go:117] "RemoveContainer" containerID="a797f4686fe9ef2de0f50ae7a0c6dc7222a7e6d08c288eee28e269c6459fc465" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.672271 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk"] Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.679112 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84bcd68b97-xz5jk"] Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.689506 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/02e920db-febc-4d97-83b4-cacd22613ebc-client-ca\") pod \"route-controller-manager-cd7bb665-nx22w\" (UID: \"02e920db-febc-4d97-83b4-cacd22613ebc\") " pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.689553 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02e920db-febc-4d97-83b4-cacd22613ebc-serving-cert\") pod \"route-controller-manager-cd7bb665-nx22w\" (UID: \"02e920db-febc-4d97-83b4-cacd22613ebc\") " pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.689598 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02e920db-febc-4d97-83b4-cacd22613ebc-config\") pod \"route-controller-manager-cd7bb665-nx22w\" (UID: \"02e920db-febc-4d97-83b4-cacd22613ebc\") " pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.689646 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgkrs\" (UniqueName: \"kubernetes.io/projected/02e920db-febc-4d97-83b4-cacd22613ebc-kube-api-access-qgkrs\") pod \"route-controller-manager-cd7bb665-nx22w\" (UID: \"02e920db-febc-4d97-83b4-cacd22613ebc\") " pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.689972 4868 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c7b3490-9561-4058-865a-7d53ca5112c5-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.689996 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c7b3490-9561-4058-865a-7d53ca5112c5-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.690008 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4kr7\" (UniqueName: \"kubernetes.io/projected/6c7b3490-9561-4058-865a-7d53ca5112c5-kube-api-access-z4kr7\") on node \"crc\" DevicePath \"\"" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.690020 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c7b3490-9561-4058-865a-7d53ca5112c5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.791083 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/02e920db-febc-4d97-83b4-cacd22613ebc-client-ca\") pod \"route-controller-manager-cd7bb665-nx22w\" (UID: \"02e920db-febc-4d97-83b4-cacd22613ebc\") " pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.791165 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02e920db-febc-4d97-83b4-cacd22613ebc-serving-cert\") pod \"route-controller-manager-cd7bb665-nx22w\" (UID: \"02e920db-febc-4d97-83b4-cacd22613ebc\") " pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.791207 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02e920db-febc-4d97-83b4-cacd22613ebc-config\") pod \"route-controller-manager-cd7bb665-nx22w\" (UID: \"02e920db-febc-4d97-83b4-cacd22613ebc\") " pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.791235 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgkrs\" (UniqueName: \"kubernetes.io/projected/02e920db-febc-4d97-83b4-cacd22613ebc-kube-api-access-qgkrs\") pod \"route-controller-manager-cd7bb665-nx22w\" (UID: \"02e920db-febc-4d97-83b4-cacd22613ebc\") " pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.792820 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/02e920db-febc-4d97-83b4-cacd22613ebc-client-ca\") pod \"route-controller-manager-cd7bb665-nx22w\" (UID: \"02e920db-febc-4d97-83b4-cacd22613ebc\") " pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.792919 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02e920db-febc-4d97-83b4-cacd22613ebc-config\") pod \"route-controller-manager-cd7bb665-nx22w\" (UID: \"02e920db-febc-4d97-83b4-cacd22613ebc\") " pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.795997 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02e920db-febc-4d97-83b4-cacd22613ebc-serving-cert\") pod \"route-controller-manager-cd7bb665-nx22w\" (UID: \"02e920db-febc-4d97-83b4-cacd22613ebc\") " pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.818199 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgkrs\" (UniqueName: \"kubernetes.io/projected/02e920db-febc-4d97-83b4-cacd22613ebc-kube-api-access-qgkrs\") pod \"route-controller-manager-cd7bb665-nx22w\" (UID: \"02e920db-febc-4d97-83b4-cacd22613ebc\") " pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:12 crc kubenswrapper[4868]: I1201 17:31:12.854399 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:13 crc kubenswrapper[4868]: I1201 17:31:13.347084 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w"] Dec 01 17:31:13 crc kubenswrapper[4868]: I1201 17:31:13.646192 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" event={"ID":"02e920db-febc-4d97-83b4-cacd22613ebc","Type":"ContainerStarted","Data":"b39884f0c8e7f61fa157c1b8fb5e5bfeb8dd36d96197b4868eff57f170850e0b"} Dec 01 17:31:13 crc kubenswrapper[4868]: I1201 17:31:13.646237 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" event={"ID":"02e920db-febc-4d97-83b4-cacd22613ebc","Type":"ContainerStarted","Data":"8d03f678c7a96d11482ee81373f5fb1abf31df090935eac53c82eb61f79be73d"} Dec 01 17:31:13 crc kubenswrapper[4868]: I1201 17:31:13.648935 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:13 crc kubenswrapper[4868]: I1201 17:31:13.667875 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" podStartSLOduration=2.667840168 podStartE2EDuration="2.667840168s" podCreationTimestamp="2025-12-01 17:31:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:31:13.663324701 +0000 UTC m=+346.034435122" watchObservedRunningTime="2025-12-01 17:31:13.667840168 +0000 UTC m=+346.038950589" Dec 01 17:31:13 crc kubenswrapper[4868]: I1201 17:31:13.897447 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-cd7bb665-nx22w" Dec 01 17:31:14 crc kubenswrapper[4868]: I1201 17:31:14.180717 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c7b3490-9561-4058-865a-7d53ca5112c5" path="/var/lib/kubelet/pods/6c7b3490-9561-4058-865a-7d53ca5112c5/volumes" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.664147 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ckkq6"] Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.665607 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.668574 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.683825 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ckkq6"] Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.746531 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/901c8b2e-6560-4d33-8233-1f01ed46b2b5-utilities\") pod \"redhat-operators-ckkq6\" (UID: \"901c8b2e-6560-4d33-8233-1f01ed46b2b5\") " pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.746603 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/901c8b2e-6560-4d33-8233-1f01ed46b2b5-catalog-content\") pod \"redhat-operators-ckkq6\" (UID: \"901c8b2e-6560-4d33-8233-1f01ed46b2b5\") " pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.746636 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz8tr\" (UniqueName: \"kubernetes.io/projected/901c8b2e-6560-4d33-8233-1f01ed46b2b5-kube-api-access-rz8tr\") pod \"redhat-operators-ckkq6\" (UID: \"901c8b2e-6560-4d33-8233-1f01ed46b2b5\") " pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.847490 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/901c8b2e-6560-4d33-8233-1f01ed46b2b5-utilities\") pod \"redhat-operators-ckkq6\" (UID: \"901c8b2e-6560-4d33-8233-1f01ed46b2b5\") " pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.847561 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/901c8b2e-6560-4d33-8233-1f01ed46b2b5-catalog-content\") pod \"redhat-operators-ckkq6\" (UID: \"901c8b2e-6560-4d33-8233-1f01ed46b2b5\") " pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.847597 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz8tr\" (UniqueName: \"kubernetes.io/projected/901c8b2e-6560-4d33-8233-1f01ed46b2b5-kube-api-access-rz8tr\") pod \"redhat-operators-ckkq6\" (UID: \"901c8b2e-6560-4d33-8233-1f01ed46b2b5\") " pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.848457 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/901c8b2e-6560-4d33-8233-1f01ed46b2b5-utilities\") pod \"redhat-operators-ckkq6\" (UID: \"901c8b2e-6560-4d33-8233-1f01ed46b2b5\") " pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.848457 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/901c8b2e-6560-4d33-8233-1f01ed46b2b5-catalog-content\") pod \"redhat-operators-ckkq6\" (UID: \"901c8b2e-6560-4d33-8233-1f01ed46b2b5\") " pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.855704 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9kf2c"] Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.857349 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9kf2c" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.859641 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.870549 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz8tr\" (UniqueName: \"kubernetes.io/projected/901c8b2e-6560-4d33-8233-1f01ed46b2b5-kube-api-access-rz8tr\") pod \"redhat-operators-ckkq6\" (UID: \"901c8b2e-6560-4d33-8233-1f01ed46b2b5\") " pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.873834 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9kf2c"] Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.949181 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bff6f35-58ce-48e1-ae66-e29973c57a3f-utilities\") pod \"redhat-marketplace-9kf2c\" (UID: \"5bff6f35-58ce-48e1-ae66-e29973c57a3f\") " pod="openshift-marketplace/redhat-marketplace-9kf2c" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.949279 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc7rs\" (UniqueName: \"kubernetes.io/projected/5bff6f35-58ce-48e1-ae66-e29973c57a3f-kube-api-access-cc7rs\") pod \"redhat-marketplace-9kf2c\" (UID: \"5bff6f35-58ce-48e1-ae66-e29973c57a3f\") " pod="openshift-marketplace/redhat-marketplace-9kf2c" Dec 01 17:31:22 crc kubenswrapper[4868]: I1201 17:31:22.949314 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bff6f35-58ce-48e1-ae66-e29973c57a3f-catalog-content\") pod \"redhat-marketplace-9kf2c\" (UID: \"5bff6f35-58ce-48e1-ae66-e29973c57a3f\") " pod="openshift-marketplace/redhat-marketplace-9kf2c" Dec 01 17:31:23 crc kubenswrapper[4868]: I1201 17:31:23.044562 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:31:23 crc kubenswrapper[4868]: I1201 17:31:23.051127 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc7rs\" (UniqueName: \"kubernetes.io/projected/5bff6f35-58ce-48e1-ae66-e29973c57a3f-kube-api-access-cc7rs\") pod \"redhat-marketplace-9kf2c\" (UID: \"5bff6f35-58ce-48e1-ae66-e29973c57a3f\") " pod="openshift-marketplace/redhat-marketplace-9kf2c" Dec 01 17:31:23 crc kubenswrapper[4868]: I1201 17:31:23.051219 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bff6f35-58ce-48e1-ae66-e29973c57a3f-catalog-content\") pod \"redhat-marketplace-9kf2c\" (UID: \"5bff6f35-58ce-48e1-ae66-e29973c57a3f\") " pod="openshift-marketplace/redhat-marketplace-9kf2c" Dec 01 17:31:23 crc kubenswrapper[4868]: I1201 17:31:23.051324 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bff6f35-58ce-48e1-ae66-e29973c57a3f-utilities\") pod \"redhat-marketplace-9kf2c\" (UID: \"5bff6f35-58ce-48e1-ae66-e29973c57a3f\") " pod="openshift-marketplace/redhat-marketplace-9kf2c" Dec 01 17:31:23 crc kubenswrapper[4868]: I1201 17:31:23.052086 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bff6f35-58ce-48e1-ae66-e29973c57a3f-catalog-content\") pod \"redhat-marketplace-9kf2c\" (UID: \"5bff6f35-58ce-48e1-ae66-e29973c57a3f\") " pod="openshift-marketplace/redhat-marketplace-9kf2c" Dec 01 17:31:23 crc kubenswrapper[4868]: I1201 17:31:23.052172 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bff6f35-58ce-48e1-ae66-e29973c57a3f-utilities\") pod \"redhat-marketplace-9kf2c\" (UID: \"5bff6f35-58ce-48e1-ae66-e29973c57a3f\") " pod="openshift-marketplace/redhat-marketplace-9kf2c" Dec 01 17:31:23 crc kubenswrapper[4868]: I1201 17:31:23.077897 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc7rs\" (UniqueName: \"kubernetes.io/projected/5bff6f35-58ce-48e1-ae66-e29973c57a3f-kube-api-access-cc7rs\") pod \"redhat-marketplace-9kf2c\" (UID: \"5bff6f35-58ce-48e1-ae66-e29973c57a3f\") " pod="openshift-marketplace/redhat-marketplace-9kf2c" Dec 01 17:31:23 crc kubenswrapper[4868]: I1201 17:31:23.195533 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9kf2c" Dec 01 17:31:23 crc kubenswrapper[4868]: I1201 17:31:23.447325 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ckkq6"] Dec 01 17:31:23 crc kubenswrapper[4868]: I1201 17:31:23.597169 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9kf2c"] Dec 01 17:31:23 crc kubenswrapper[4868]: W1201 17:31:23.613422 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5bff6f35_58ce_48e1_ae66_e29973c57a3f.slice/crio-3b64ac036932bbcf271d7ac1e8271bbd0cf89ccb824ab84e7a3c28189ee56b14 WatchSource:0}: Error finding container 3b64ac036932bbcf271d7ac1e8271bbd0cf89ccb824ab84e7a3c28189ee56b14: Status 404 returned error can't find the container with id 3b64ac036932bbcf271d7ac1e8271bbd0cf89ccb824ab84e7a3c28189ee56b14 Dec 01 17:31:23 crc kubenswrapper[4868]: I1201 17:31:23.746349 4868 generic.go:334] "Generic (PLEG): container finished" podID="901c8b2e-6560-4d33-8233-1f01ed46b2b5" containerID="fb4a9edadc208dba4a5b062181a3005ed052eb60cf5beed34bd882c777762e02" exitCode=0 Dec 01 17:31:23 crc kubenswrapper[4868]: I1201 17:31:23.746446 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ckkq6" event={"ID":"901c8b2e-6560-4d33-8233-1f01ed46b2b5","Type":"ContainerDied","Data":"fb4a9edadc208dba4a5b062181a3005ed052eb60cf5beed34bd882c777762e02"} Dec 01 17:31:23 crc kubenswrapper[4868]: I1201 17:31:23.746488 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ckkq6" event={"ID":"901c8b2e-6560-4d33-8233-1f01ed46b2b5","Type":"ContainerStarted","Data":"86678aec97a45d02dff866087ac60031e47bf279656436581fa8f819db096f9b"} Dec 01 17:31:23 crc kubenswrapper[4868]: I1201 17:31:23.747505 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kf2c" event={"ID":"5bff6f35-58ce-48e1-ae66-e29973c57a3f","Type":"ContainerStarted","Data":"3b64ac036932bbcf271d7ac1e8271bbd0cf89ccb824ab84e7a3c28189ee56b14"} Dec 01 17:31:24 crc kubenswrapper[4868]: I1201 17:31:24.770189 4868 generic.go:334] "Generic (PLEG): container finished" podID="5bff6f35-58ce-48e1-ae66-e29973c57a3f" containerID="6c52d6b565f92b5c11bd4f24f51ea96c41491999dc517b2c2ab68f704a8b74ba" exitCode=0 Dec 01 17:31:24 crc kubenswrapper[4868]: I1201 17:31:24.770324 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kf2c" event={"ID":"5bff6f35-58ce-48e1-ae66-e29973c57a3f","Type":"ContainerDied","Data":"6c52d6b565f92b5c11bd4f24f51ea96c41491999dc517b2c2ab68f704a8b74ba"} Dec 01 17:31:24 crc kubenswrapper[4868]: I1201 17:31:24.777365 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ckkq6" event={"ID":"901c8b2e-6560-4d33-8233-1f01ed46b2b5","Type":"ContainerStarted","Data":"8f2fe03da4a9272ef15092e2794d94c916a1c8cdd19cc20da81456ff942d7ae3"} Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.055462 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-znxj6"] Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.056661 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znxj6" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.060176 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.067265 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-znxj6"] Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.204079 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwz6x\" (UniqueName: \"kubernetes.io/projected/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-kube-api-access-pwz6x\") pod \"certified-operators-znxj6\" (UID: \"2faafa31-ca8e-4da4-a1b4-e3d2709b4648\") " pod="openshift-marketplace/certified-operators-znxj6" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.204120 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-catalog-content\") pod \"certified-operators-znxj6\" (UID: \"2faafa31-ca8e-4da4-a1b4-e3d2709b4648\") " pod="openshift-marketplace/certified-operators-znxj6" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.204227 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-utilities\") pod \"certified-operators-znxj6\" (UID: \"2faafa31-ca8e-4da4-a1b4-e3d2709b4648\") " pod="openshift-marketplace/certified-operators-znxj6" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.253616 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2b5vb"] Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.255168 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2b5vb" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.267092 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2b5vb"] Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.267932 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.305910 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwz6x\" (UniqueName: \"kubernetes.io/projected/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-kube-api-access-pwz6x\") pod \"certified-operators-znxj6\" (UID: \"2faafa31-ca8e-4da4-a1b4-e3d2709b4648\") " pod="openshift-marketplace/certified-operators-znxj6" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.305997 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-catalog-content\") pod \"certified-operators-znxj6\" (UID: \"2faafa31-ca8e-4da4-a1b4-e3d2709b4648\") " pod="openshift-marketplace/certified-operators-znxj6" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.306060 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-utilities\") pod \"certified-operators-znxj6\" (UID: \"2faafa31-ca8e-4da4-a1b4-e3d2709b4648\") " pod="openshift-marketplace/certified-operators-znxj6" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.306556 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-catalog-content\") pod \"certified-operators-znxj6\" (UID: \"2faafa31-ca8e-4da4-a1b4-e3d2709b4648\") " pod="openshift-marketplace/certified-operators-znxj6" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.306649 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-utilities\") pod \"certified-operators-znxj6\" (UID: \"2faafa31-ca8e-4da4-a1b4-e3d2709b4648\") " pod="openshift-marketplace/certified-operators-znxj6" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.334196 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwz6x\" (UniqueName: \"kubernetes.io/projected/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-kube-api-access-pwz6x\") pod \"certified-operators-znxj6\" (UID: \"2faafa31-ca8e-4da4-a1b4-e3d2709b4648\") " pod="openshift-marketplace/certified-operators-znxj6" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.407161 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05d6020b-0574-40f5-b421-fe6b21712fd2-utilities\") pod \"community-operators-2b5vb\" (UID: \"05d6020b-0574-40f5-b421-fe6b21712fd2\") " pod="openshift-marketplace/community-operators-2b5vb" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.407350 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrjc2\" (UniqueName: \"kubernetes.io/projected/05d6020b-0574-40f5-b421-fe6b21712fd2-kube-api-access-hrjc2\") pod \"community-operators-2b5vb\" (UID: \"05d6020b-0574-40f5-b421-fe6b21712fd2\") " pod="openshift-marketplace/community-operators-2b5vb" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.407524 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05d6020b-0574-40f5-b421-fe6b21712fd2-catalog-content\") pod \"community-operators-2b5vb\" (UID: \"05d6020b-0574-40f5-b421-fe6b21712fd2\") " pod="openshift-marketplace/community-operators-2b5vb" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.417815 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znxj6" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.509440 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05d6020b-0574-40f5-b421-fe6b21712fd2-utilities\") pod \"community-operators-2b5vb\" (UID: \"05d6020b-0574-40f5-b421-fe6b21712fd2\") " pod="openshift-marketplace/community-operators-2b5vb" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.509813 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrjc2\" (UniqueName: \"kubernetes.io/projected/05d6020b-0574-40f5-b421-fe6b21712fd2-kube-api-access-hrjc2\") pod \"community-operators-2b5vb\" (UID: \"05d6020b-0574-40f5-b421-fe6b21712fd2\") " pod="openshift-marketplace/community-operators-2b5vb" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.509867 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05d6020b-0574-40f5-b421-fe6b21712fd2-catalog-content\") pod \"community-operators-2b5vb\" (UID: \"05d6020b-0574-40f5-b421-fe6b21712fd2\") " pod="openshift-marketplace/community-operators-2b5vb" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.510696 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05d6020b-0574-40f5-b421-fe6b21712fd2-catalog-content\") pod \"community-operators-2b5vb\" (UID: \"05d6020b-0574-40f5-b421-fe6b21712fd2\") " pod="openshift-marketplace/community-operators-2b5vb" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.510916 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05d6020b-0574-40f5-b421-fe6b21712fd2-utilities\") pod \"community-operators-2b5vb\" (UID: \"05d6020b-0574-40f5-b421-fe6b21712fd2\") " pod="openshift-marketplace/community-operators-2b5vb" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.532099 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrjc2\" (UniqueName: \"kubernetes.io/projected/05d6020b-0574-40f5-b421-fe6b21712fd2-kube-api-access-hrjc2\") pod \"community-operators-2b5vb\" (UID: \"05d6020b-0574-40f5-b421-fe6b21712fd2\") " pod="openshift-marketplace/community-operators-2b5vb" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.570701 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2b5vb" Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.785599 4868 generic.go:334] "Generic (PLEG): container finished" podID="901c8b2e-6560-4d33-8233-1f01ed46b2b5" containerID="8f2fe03da4a9272ef15092e2794d94c916a1c8cdd19cc20da81456ff942d7ae3" exitCode=0 Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.785677 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ckkq6" event={"ID":"901c8b2e-6560-4d33-8233-1f01ed46b2b5","Type":"ContainerDied","Data":"8f2fe03da4a9272ef15092e2794d94c916a1c8cdd19cc20da81456ff942d7ae3"} Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.789870 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kf2c" event={"ID":"5bff6f35-58ce-48e1-ae66-e29973c57a3f","Type":"ContainerStarted","Data":"13838d9f8ef0c8aa3e7ed0c46f9d782d079dca167a2f3fa3604c531fe4395e73"} Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.844350 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-znxj6"] Dec 01 17:31:25 crc kubenswrapper[4868]: W1201 17:31:25.849764 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2faafa31_ca8e_4da4_a1b4_e3d2709b4648.slice/crio-1115711895dbf0833014baf2f5523dff2a5a1ccba8954e931b7a87d1fddfeec9 WatchSource:0}: Error finding container 1115711895dbf0833014baf2f5523dff2a5a1ccba8954e931b7a87d1fddfeec9: Status 404 returned error can't find the container with id 1115711895dbf0833014baf2f5523dff2a5a1ccba8954e931b7a87d1fddfeec9 Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.906111 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:31:25 crc kubenswrapper[4868]: I1201 17:31:25.906209 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:31:26 crc kubenswrapper[4868]: I1201 17:31:26.019081 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2b5vb"] Dec 01 17:31:26 crc kubenswrapper[4868]: W1201 17:31:26.025934 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05d6020b_0574_40f5_b421_fe6b21712fd2.slice/crio-4ab5b92cd7ad6e093f6dc67f3f54c0370389437bc68bd5adc53f3fc4c63c6906 WatchSource:0}: Error finding container 4ab5b92cd7ad6e093f6dc67f3f54c0370389437bc68bd5adc53f3fc4c63c6906: Status 404 returned error can't find the container with id 4ab5b92cd7ad6e093f6dc67f3f54c0370389437bc68bd5adc53f3fc4c63c6906 Dec 01 17:31:26 crc kubenswrapper[4868]: I1201 17:31:26.801128 4868 generic.go:334] "Generic (PLEG): container finished" podID="5bff6f35-58ce-48e1-ae66-e29973c57a3f" containerID="13838d9f8ef0c8aa3e7ed0c46f9d782d079dca167a2f3fa3604c531fe4395e73" exitCode=0 Dec 01 17:31:26 crc kubenswrapper[4868]: I1201 17:31:26.801748 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kf2c" event={"ID":"5bff6f35-58ce-48e1-ae66-e29973c57a3f","Type":"ContainerDied","Data":"13838d9f8ef0c8aa3e7ed0c46f9d782d079dca167a2f3fa3604c531fe4395e73"} Dec 01 17:31:26 crc kubenswrapper[4868]: I1201 17:31:26.804283 4868 generic.go:334] "Generic (PLEG): container finished" podID="05d6020b-0574-40f5-b421-fe6b21712fd2" containerID="0c905699ffbd48633864afcc85b9b2c2159ade7d0493a9377771f6382d758587" exitCode=0 Dec 01 17:31:26 crc kubenswrapper[4868]: I1201 17:31:26.804345 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2b5vb" event={"ID":"05d6020b-0574-40f5-b421-fe6b21712fd2","Type":"ContainerDied","Data":"0c905699ffbd48633864afcc85b9b2c2159ade7d0493a9377771f6382d758587"} Dec 01 17:31:26 crc kubenswrapper[4868]: I1201 17:31:26.804371 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2b5vb" event={"ID":"05d6020b-0574-40f5-b421-fe6b21712fd2","Type":"ContainerStarted","Data":"4ab5b92cd7ad6e093f6dc67f3f54c0370389437bc68bd5adc53f3fc4c63c6906"} Dec 01 17:31:26 crc kubenswrapper[4868]: I1201 17:31:26.808290 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ckkq6" event={"ID":"901c8b2e-6560-4d33-8233-1f01ed46b2b5","Type":"ContainerStarted","Data":"97dc7423675eb5f575f9660a1f97eeeb6a9090124b404f56eef1bd8461f793bf"} Dec 01 17:31:26 crc kubenswrapper[4868]: I1201 17:31:26.810989 4868 generic.go:334] "Generic (PLEG): container finished" podID="2faafa31-ca8e-4da4-a1b4-e3d2709b4648" containerID="80079058bbb927e2310b6503faea9c0bccef225b615cbd38b8fdf5595d3b641d" exitCode=0 Dec 01 17:31:26 crc kubenswrapper[4868]: I1201 17:31:26.811063 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znxj6" event={"ID":"2faafa31-ca8e-4da4-a1b4-e3d2709b4648","Type":"ContainerDied","Data":"80079058bbb927e2310b6503faea9c0bccef225b615cbd38b8fdf5595d3b641d"} Dec 01 17:31:26 crc kubenswrapper[4868]: I1201 17:31:26.811104 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znxj6" event={"ID":"2faafa31-ca8e-4da4-a1b4-e3d2709b4648","Type":"ContainerStarted","Data":"1115711895dbf0833014baf2f5523dff2a5a1ccba8954e931b7a87d1fddfeec9"} Dec 01 17:31:26 crc kubenswrapper[4868]: I1201 17:31:26.878251 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ckkq6" podStartSLOduration=2.251641815 podStartE2EDuration="4.878224457s" podCreationTimestamp="2025-12-01 17:31:22 +0000 UTC" firstStartedPulling="2025-12-01 17:31:23.74894938 +0000 UTC m=+356.120059791" lastFinishedPulling="2025-12-01 17:31:26.375532022 +0000 UTC m=+358.746642433" observedRunningTime="2025-12-01 17:31:26.873602516 +0000 UTC m=+359.244712957" watchObservedRunningTime="2025-12-01 17:31:26.878224457 +0000 UTC m=+359.249334878" Dec 01 17:31:27 crc kubenswrapper[4868]: I1201 17:31:27.819187 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kf2c" event={"ID":"5bff6f35-58ce-48e1-ae66-e29973c57a3f","Type":"ContainerStarted","Data":"646f02843c9f328729aefb948b91e3cc140ff04d783bd2cc0db19d0c119d2bdb"} Dec 01 17:31:27 crc kubenswrapper[4868]: I1201 17:31:27.820767 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znxj6" event={"ID":"2faafa31-ca8e-4da4-a1b4-e3d2709b4648","Type":"ContainerStarted","Data":"a013d6acaa5ac65f3bba8db55766eb4a784131fb28e4b1fc0da1e74b201da117"} Dec 01 17:31:27 crc kubenswrapper[4868]: I1201 17:31:27.845916 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9kf2c" podStartSLOduration=3.160608308 podStartE2EDuration="5.845883051s" podCreationTimestamp="2025-12-01 17:31:22 +0000 UTC" firstStartedPulling="2025-12-01 17:31:24.772767833 +0000 UTC m=+357.143878244" lastFinishedPulling="2025-12-01 17:31:27.458042566 +0000 UTC m=+359.829152987" observedRunningTime="2025-12-01 17:31:27.842922057 +0000 UTC m=+360.214032498" watchObservedRunningTime="2025-12-01 17:31:27.845883051 +0000 UTC m=+360.216993502" Dec 01 17:31:28 crc kubenswrapper[4868]: I1201 17:31:28.830564 4868 generic.go:334] "Generic (PLEG): container finished" podID="05d6020b-0574-40f5-b421-fe6b21712fd2" containerID="c9e172b052a16138cfb8346293175e6ab32cbcc88b89d1be33a63fc0e8cbc65a" exitCode=0 Dec 01 17:31:28 crc kubenswrapper[4868]: I1201 17:31:28.830679 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2b5vb" event={"ID":"05d6020b-0574-40f5-b421-fe6b21712fd2","Type":"ContainerDied","Data":"c9e172b052a16138cfb8346293175e6ab32cbcc88b89d1be33a63fc0e8cbc65a"} Dec 01 17:31:28 crc kubenswrapper[4868]: I1201 17:31:28.838361 4868 generic.go:334] "Generic (PLEG): container finished" podID="2faafa31-ca8e-4da4-a1b4-e3d2709b4648" containerID="a013d6acaa5ac65f3bba8db55766eb4a784131fb28e4b1fc0da1e74b201da117" exitCode=0 Dec 01 17:31:28 crc kubenswrapper[4868]: I1201 17:31:28.839564 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znxj6" event={"ID":"2faafa31-ca8e-4da4-a1b4-e3d2709b4648","Type":"ContainerDied","Data":"a013d6acaa5ac65f3bba8db55766eb4a784131fb28e4b1fc0da1e74b201da117"} Dec 01 17:31:29 crc kubenswrapper[4868]: I1201 17:31:29.846768 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2b5vb" event={"ID":"05d6020b-0574-40f5-b421-fe6b21712fd2","Type":"ContainerStarted","Data":"721add12e4b66e4d21eb3e6db6f32a0316deb250a95a47c117ef73d74077d7c9"} Dec 01 17:31:29 crc kubenswrapper[4868]: I1201 17:31:29.858118 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znxj6" event={"ID":"2faafa31-ca8e-4da4-a1b4-e3d2709b4648","Type":"ContainerStarted","Data":"6535c775409e77f58f8df53032a1d000c55894a88d6d0e340e911f758a67c8ca"} Dec 01 17:31:29 crc kubenswrapper[4868]: I1201 17:31:29.880905 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2b5vb" podStartSLOduration=2.157360715 podStartE2EDuration="4.880882511s" podCreationTimestamp="2025-12-01 17:31:25 +0000 UTC" firstStartedPulling="2025-12-01 17:31:26.809517952 +0000 UTC m=+359.180628373" lastFinishedPulling="2025-12-01 17:31:29.533039758 +0000 UTC m=+361.904150169" observedRunningTime="2025-12-01 17:31:29.87728854 +0000 UTC m=+362.248398951" watchObservedRunningTime="2025-12-01 17:31:29.880882511 +0000 UTC m=+362.251992922" Dec 01 17:31:29 crc kubenswrapper[4868]: I1201 17:31:29.925538 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-znxj6" podStartSLOduration=2.293443998 podStartE2EDuration="4.925516244s" podCreationTimestamp="2025-12-01 17:31:25 +0000 UTC" firstStartedPulling="2025-12-01 17:31:26.812737644 +0000 UTC m=+359.183848065" lastFinishedPulling="2025-12-01 17:31:29.4448099 +0000 UTC m=+361.815920311" observedRunningTime="2025-12-01 17:31:29.923572549 +0000 UTC m=+362.294682970" watchObservedRunningTime="2025-12-01 17:31:29.925516244 +0000 UTC m=+362.296626655" Dec 01 17:31:33 crc kubenswrapper[4868]: I1201 17:31:33.045397 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:31:33 crc kubenswrapper[4868]: I1201 17:31:33.045731 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:31:33 crc kubenswrapper[4868]: I1201 17:31:33.086722 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:31:33 crc kubenswrapper[4868]: I1201 17:31:33.197150 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9kf2c" Dec 01 17:31:33 crc kubenswrapper[4868]: I1201 17:31:33.198191 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9kf2c" Dec 01 17:31:33 crc kubenswrapper[4868]: I1201 17:31:33.236476 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9kf2c" Dec 01 17:31:33 crc kubenswrapper[4868]: I1201 17:31:33.933854 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9kf2c" Dec 01 17:31:33 crc kubenswrapper[4868]: I1201 17:31:33.934225 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:31:35 crc kubenswrapper[4868]: I1201 17:31:35.418245 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-znxj6" Dec 01 17:31:35 crc kubenswrapper[4868]: I1201 17:31:35.418770 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-znxj6" Dec 01 17:31:35 crc kubenswrapper[4868]: I1201 17:31:35.470683 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-znxj6" Dec 01 17:31:35 crc kubenswrapper[4868]: I1201 17:31:35.572553 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2b5vb" Dec 01 17:31:35 crc kubenswrapper[4868]: I1201 17:31:35.572667 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2b5vb" Dec 01 17:31:35 crc kubenswrapper[4868]: I1201 17:31:35.626464 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2b5vb" Dec 01 17:31:35 crc kubenswrapper[4868]: I1201 17:31:35.934453 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-znxj6" Dec 01 17:31:35 crc kubenswrapper[4868]: I1201 17:31:35.937653 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2b5vb" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.400135 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tk8hc"] Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.401607 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.428517 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tk8hc"] Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.562817 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ec3e82f7-d487-4023-b049-7e55baf7cde8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.562884 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ec3e82f7-d487-4023-b049-7e55baf7cde8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.562922 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ec3e82f7-d487-4023-b049-7e55baf7cde8-bound-sa-token\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.562967 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ec3e82f7-d487-4023-b049-7e55baf7cde8-registry-certificates\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.563004 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ec3e82f7-d487-4023-b049-7e55baf7cde8-trusted-ca\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.563023 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prsrj\" (UniqueName: \"kubernetes.io/projected/ec3e82f7-d487-4023-b049-7e55baf7cde8-kube-api-access-prsrj\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.563067 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.563118 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ec3e82f7-d487-4023-b049-7e55baf7cde8-registry-tls\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.593721 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.665195 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ec3e82f7-d487-4023-b049-7e55baf7cde8-registry-tls\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.665654 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ec3e82f7-d487-4023-b049-7e55baf7cde8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.665753 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ec3e82f7-d487-4023-b049-7e55baf7cde8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.665868 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ec3e82f7-d487-4023-b049-7e55baf7cde8-bound-sa-token\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.665980 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ec3e82f7-d487-4023-b049-7e55baf7cde8-registry-certificates\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.666098 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ec3e82f7-d487-4023-b049-7e55baf7cde8-trusted-ca\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.666175 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prsrj\" (UniqueName: \"kubernetes.io/projected/ec3e82f7-d487-4023-b049-7e55baf7cde8-kube-api-access-prsrj\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.666242 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ec3e82f7-d487-4023-b049-7e55baf7cde8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.667139 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ec3e82f7-d487-4023-b049-7e55baf7cde8-trusted-ca\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.668115 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ec3e82f7-d487-4023-b049-7e55baf7cde8-registry-certificates\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.672413 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ec3e82f7-d487-4023-b049-7e55baf7cde8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.673893 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ec3e82f7-d487-4023-b049-7e55baf7cde8-registry-tls\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.690561 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prsrj\" (UniqueName: \"kubernetes.io/projected/ec3e82f7-d487-4023-b049-7e55baf7cde8-kube-api-access-prsrj\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.696538 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ec3e82f7-d487-4023-b049-7e55baf7cde8-bound-sa-token\") pod \"image-registry-66df7c8f76-tk8hc\" (UID: \"ec3e82f7-d487-4023-b049-7e55baf7cde8\") " pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:41 crc kubenswrapper[4868]: I1201 17:31:41.720404 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:42 crc kubenswrapper[4868]: I1201 17:31:42.178590 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tk8hc"] Dec 01 17:31:42 crc kubenswrapper[4868]: I1201 17:31:42.943823 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" event={"ID":"ec3e82f7-d487-4023-b049-7e55baf7cde8","Type":"ContainerStarted","Data":"b4c305aadb4e2e4ecac4361da0565a6712fec46af8b3bc62b741dd84807dfd8f"} Dec 01 17:31:42 crc kubenswrapper[4868]: I1201 17:31:42.945760 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" event={"ID":"ec3e82f7-d487-4023-b049-7e55baf7cde8","Type":"ContainerStarted","Data":"320fc2c063594265e3e5be135919342f70d8852de1adfc19ff03c0a1a094c6f3"} Dec 01 17:31:42 crc kubenswrapper[4868]: I1201 17:31:42.945916 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:31:42 crc kubenswrapper[4868]: I1201 17:31:42.973639 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" podStartSLOduration=1.973607348 podStartE2EDuration="1.973607348s" podCreationTimestamp="2025-12-01 17:31:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:31:42.969559294 +0000 UTC m=+375.340669705" watchObservedRunningTime="2025-12-01 17:31:42.973607348 +0000 UTC m=+375.344717759" Dec 01 17:31:51 crc kubenswrapper[4868]: I1201 17:31:51.480436 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-599f95c79c-h76w9"] Dec 01 17:31:51 crc kubenswrapper[4868]: I1201 17:31:51.481264 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" podUID="26db4f43-c152-44a3-be40-c00407d62fb0" containerName="controller-manager" containerID="cri-o://0014d7a143cbbc6afe75079e23a3ef9427b4d11f21253762f4d29aef535a61ab" gracePeriod=30 Dec 01 17:31:51 crc kubenswrapper[4868]: I1201 17:31:51.891227 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.014157 4868 generic.go:334] "Generic (PLEG): container finished" podID="26db4f43-c152-44a3-be40-c00407d62fb0" containerID="0014d7a143cbbc6afe75079e23a3ef9427b4d11f21253762f4d29aef535a61ab" exitCode=0 Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.014239 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" event={"ID":"26db4f43-c152-44a3-be40-c00407d62fb0","Type":"ContainerDied","Data":"0014d7a143cbbc6afe75079e23a3ef9427b4d11f21253762f4d29aef535a61ab"} Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.014272 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.014322 4868 scope.go:117] "RemoveContainer" containerID="0014d7a143cbbc6afe75079e23a3ef9427b4d11f21253762f4d29aef535a61ab" Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.014301 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-599f95c79c-h76w9" event={"ID":"26db4f43-c152-44a3-be40-c00407d62fb0","Type":"ContainerDied","Data":"a3d8b15adb2848467cfe46231c22e740519b0196d8b5b69344bbe68d5ed937ad"} Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.039981 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqcsg\" (UniqueName: \"kubernetes.io/projected/26db4f43-c152-44a3-be40-c00407d62fb0-kube-api-access-wqcsg\") pod \"26db4f43-c152-44a3-be40-c00407d62fb0\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.040106 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-config\") pod \"26db4f43-c152-44a3-be40-c00407d62fb0\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.040151 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-proxy-ca-bundles\") pod \"26db4f43-c152-44a3-be40-c00407d62fb0\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.040182 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-client-ca\") pod \"26db4f43-c152-44a3-be40-c00407d62fb0\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.040240 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26db4f43-c152-44a3-be40-c00407d62fb0-serving-cert\") pod \"26db4f43-c152-44a3-be40-c00407d62fb0\" (UID: \"26db4f43-c152-44a3-be40-c00407d62fb0\") " Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.041258 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "26db4f43-c152-44a3-be40-c00407d62fb0" (UID: "26db4f43-c152-44a3-be40-c00407d62fb0"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.041313 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-client-ca" (OuterVolumeSpecName: "client-ca") pod "26db4f43-c152-44a3-be40-c00407d62fb0" (UID: "26db4f43-c152-44a3-be40-c00407d62fb0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.041387 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-config" (OuterVolumeSpecName: "config") pod "26db4f43-c152-44a3-be40-c00407d62fb0" (UID: "26db4f43-c152-44a3-be40-c00407d62fb0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.042734 4868 scope.go:117] "RemoveContainer" containerID="0014d7a143cbbc6afe75079e23a3ef9427b4d11f21253762f4d29aef535a61ab" Dec 01 17:31:52 crc kubenswrapper[4868]: E1201 17:31:52.043563 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0014d7a143cbbc6afe75079e23a3ef9427b4d11f21253762f4d29aef535a61ab\": container with ID starting with 0014d7a143cbbc6afe75079e23a3ef9427b4d11f21253762f4d29aef535a61ab not found: ID does not exist" containerID="0014d7a143cbbc6afe75079e23a3ef9427b4d11f21253762f4d29aef535a61ab" Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.043637 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0014d7a143cbbc6afe75079e23a3ef9427b4d11f21253762f4d29aef535a61ab"} err="failed to get container status \"0014d7a143cbbc6afe75079e23a3ef9427b4d11f21253762f4d29aef535a61ab\": rpc error: code = NotFound desc = could not find container \"0014d7a143cbbc6afe75079e23a3ef9427b4d11f21253762f4d29aef535a61ab\": container with ID starting with 0014d7a143cbbc6afe75079e23a3ef9427b4d11f21253762f4d29aef535a61ab not found: ID does not exist" Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.058963 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26db4f43-c152-44a3-be40-c00407d62fb0-kube-api-access-wqcsg" (OuterVolumeSpecName: "kube-api-access-wqcsg") pod "26db4f43-c152-44a3-be40-c00407d62fb0" (UID: "26db4f43-c152-44a3-be40-c00407d62fb0"). InnerVolumeSpecName "kube-api-access-wqcsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.058962 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26db4f43-c152-44a3-be40-c00407d62fb0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "26db4f43-c152-44a3-be40-c00407d62fb0" (UID: "26db4f43-c152-44a3-be40-c00407d62fb0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.142332 4868 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26db4f43-c152-44a3-be40-c00407d62fb0-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.142381 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqcsg\" (UniqueName: \"kubernetes.io/projected/26db4f43-c152-44a3-be40-c00407d62fb0-kube-api-access-wqcsg\") on node \"crc\" DevicePath \"\"" Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.142394 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.142402 4868 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.142412 4868 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/26db4f43-c152-44a3-be40-c00407d62fb0-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.351987 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-599f95c79c-h76w9"] Dec 01 17:31:52 crc kubenswrapper[4868]: I1201 17:31:52.358004 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-599f95c79c-h76w9"] Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.067678 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf"] Dec 01 17:31:53 crc kubenswrapper[4868]: E1201 17:31:53.068763 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26db4f43-c152-44a3-be40-c00407d62fb0" containerName="controller-manager" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.068782 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="26db4f43-c152-44a3-be40-c00407d62fb0" containerName="controller-manager" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.068976 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="26db4f43-c152-44a3-be40-c00407d62fb0" containerName="controller-manager" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.070247 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.074578 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.075375 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.077225 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.077599 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.083153 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.083250 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.089664 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.094766 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf"] Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.166283 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cd64dcd4-3b89-4662-86db-209dba199c5f-client-ca\") pod \"controller-manager-54d8f4c94f-c2hdf\" (UID: \"cd64dcd4-3b89-4662-86db-209dba199c5f\") " pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.166339 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cd64dcd4-3b89-4662-86db-209dba199c5f-proxy-ca-bundles\") pod \"controller-manager-54d8f4c94f-c2hdf\" (UID: \"cd64dcd4-3b89-4662-86db-209dba199c5f\") " pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.166386 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd64dcd4-3b89-4662-86db-209dba199c5f-config\") pod \"controller-manager-54d8f4c94f-c2hdf\" (UID: \"cd64dcd4-3b89-4662-86db-209dba199c5f\") " pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.166409 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgsh5\" (UniqueName: \"kubernetes.io/projected/cd64dcd4-3b89-4662-86db-209dba199c5f-kube-api-access-rgsh5\") pod \"controller-manager-54d8f4c94f-c2hdf\" (UID: \"cd64dcd4-3b89-4662-86db-209dba199c5f\") " pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.166441 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd64dcd4-3b89-4662-86db-209dba199c5f-serving-cert\") pod \"controller-manager-54d8f4c94f-c2hdf\" (UID: \"cd64dcd4-3b89-4662-86db-209dba199c5f\") " pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.268093 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cd64dcd4-3b89-4662-86db-209dba199c5f-client-ca\") pod \"controller-manager-54d8f4c94f-c2hdf\" (UID: \"cd64dcd4-3b89-4662-86db-209dba199c5f\") " pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.268157 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cd64dcd4-3b89-4662-86db-209dba199c5f-proxy-ca-bundles\") pod \"controller-manager-54d8f4c94f-c2hdf\" (UID: \"cd64dcd4-3b89-4662-86db-209dba199c5f\") " pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.268227 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd64dcd4-3b89-4662-86db-209dba199c5f-config\") pod \"controller-manager-54d8f4c94f-c2hdf\" (UID: \"cd64dcd4-3b89-4662-86db-209dba199c5f\") " pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.268251 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgsh5\" (UniqueName: \"kubernetes.io/projected/cd64dcd4-3b89-4662-86db-209dba199c5f-kube-api-access-rgsh5\") pod \"controller-manager-54d8f4c94f-c2hdf\" (UID: \"cd64dcd4-3b89-4662-86db-209dba199c5f\") " pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.268292 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd64dcd4-3b89-4662-86db-209dba199c5f-serving-cert\") pod \"controller-manager-54d8f4c94f-c2hdf\" (UID: \"cd64dcd4-3b89-4662-86db-209dba199c5f\") " pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.270544 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cd64dcd4-3b89-4662-86db-209dba199c5f-client-ca\") pod \"controller-manager-54d8f4c94f-c2hdf\" (UID: \"cd64dcd4-3b89-4662-86db-209dba199c5f\") " pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.273002 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cd64dcd4-3b89-4662-86db-209dba199c5f-proxy-ca-bundles\") pod \"controller-manager-54d8f4c94f-c2hdf\" (UID: \"cd64dcd4-3b89-4662-86db-209dba199c5f\") " pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.273739 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd64dcd4-3b89-4662-86db-209dba199c5f-config\") pod \"controller-manager-54d8f4c94f-c2hdf\" (UID: \"cd64dcd4-3b89-4662-86db-209dba199c5f\") " pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.275462 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd64dcd4-3b89-4662-86db-209dba199c5f-serving-cert\") pod \"controller-manager-54d8f4c94f-c2hdf\" (UID: \"cd64dcd4-3b89-4662-86db-209dba199c5f\") " pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.287793 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgsh5\" (UniqueName: \"kubernetes.io/projected/cd64dcd4-3b89-4662-86db-209dba199c5f-kube-api-access-rgsh5\") pod \"controller-manager-54d8f4c94f-c2hdf\" (UID: \"cd64dcd4-3b89-4662-86db-209dba199c5f\") " pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.410461 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:53 crc kubenswrapper[4868]: I1201 17:31:53.882962 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf"] Dec 01 17:31:53 crc kubenswrapper[4868]: W1201 17:31:53.893164 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd64dcd4_3b89_4662_86db_209dba199c5f.slice/crio-29c45cfe80498fa3848b122682b6b13915f18efed10cce1cda32c0274938d6fb WatchSource:0}: Error finding container 29c45cfe80498fa3848b122682b6b13915f18efed10cce1cda32c0274938d6fb: Status 404 returned error can't find the container with id 29c45cfe80498fa3848b122682b6b13915f18efed10cce1cda32c0274938d6fb Dec 01 17:31:54 crc kubenswrapper[4868]: I1201 17:31:54.033721 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" event={"ID":"cd64dcd4-3b89-4662-86db-209dba199c5f","Type":"ContainerStarted","Data":"762c1bb45d563b3be0f89ba3172a49f47eea35e572526e0f61c280d9421c9ca2"} Dec 01 17:31:54 crc kubenswrapper[4868]: I1201 17:31:54.033780 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" event={"ID":"cd64dcd4-3b89-4662-86db-209dba199c5f","Type":"ContainerStarted","Data":"29c45cfe80498fa3848b122682b6b13915f18efed10cce1cda32c0274938d6fb"} Dec 01 17:31:54 crc kubenswrapper[4868]: I1201 17:31:54.034717 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:54 crc kubenswrapper[4868]: I1201 17:31:54.036466 4868 patch_prober.go:28] interesting pod/controller-manager-54d8f4c94f-c2hdf container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" start-of-body= Dec 01 17:31:54 crc kubenswrapper[4868]: I1201 17:31:54.036511 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" podUID="cd64dcd4-3b89-4662-86db-209dba199c5f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" Dec 01 17:31:54 crc kubenswrapper[4868]: I1201 17:31:54.063911 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" podStartSLOduration=3.06387441 podStartE2EDuration="3.06387441s" podCreationTimestamp="2025-12-01 17:31:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:31:54.056481262 +0000 UTC m=+386.427591703" watchObservedRunningTime="2025-12-01 17:31:54.06387441 +0000 UTC m=+386.434984831" Dec 01 17:31:54 crc kubenswrapper[4868]: I1201 17:31:54.186535 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26db4f43-c152-44a3-be40-c00407d62fb0" path="/var/lib/kubelet/pods/26db4f43-c152-44a3-be40-c00407d62fb0/volumes" Dec 01 17:31:55 crc kubenswrapper[4868]: I1201 17:31:55.047653 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-54d8f4c94f-c2hdf" Dec 01 17:31:55 crc kubenswrapper[4868]: I1201 17:31:55.904976 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:31:55 crc kubenswrapper[4868]: I1201 17:31:55.905518 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:32:01 crc kubenswrapper[4868]: I1201 17:32:01.734214 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-tk8hc" Dec 01 17:32:01 crc kubenswrapper[4868]: I1201 17:32:01.798964 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-x5gxr"] Dec 01 17:32:25 crc kubenswrapper[4868]: I1201 17:32:25.904973 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:32:25 crc kubenswrapper[4868]: I1201 17:32:25.905814 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:32:25 crc kubenswrapper[4868]: I1201 17:32:25.905905 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:32:25 crc kubenswrapper[4868]: I1201 17:32:25.906728 4868 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5a50633c9f5f1a502388b67f58e02c5f45fd0e84eeadf70b339824d2940a6594"} pod="openshift-machine-config-operator/machine-config-daemon-njgpn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 17:32:25 crc kubenswrapper[4868]: I1201 17:32:25.906807 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" containerID="cri-o://5a50633c9f5f1a502388b67f58e02c5f45fd0e84eeadf70b339824d2940a6594" gracePeriod=600 Dec 01 17:32:26 crc kubenswrapper[4868]: I1201 17:32:26.253293 4868 generic.go:334] "Generic (PLEG): container finished" podID="ec4e08bf-df56-4490-8e79-a4175c727405" containerID="5a50633c9f5f1a502388b67f58e02c5f45fd0e84eeadf70b339824d2940a6594" exitCode=0 Dec 01 17:32:26 crc kubenswrapper[4868]: I1201 17:32:26.253365 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerDied","Data":"5a50633c9f5f1a502388b67f58e02c5f45fd0e84eeadf70b339824d2940a6594"} Dec 01 17:32:26 crc kubenswrapper[4868]: I1201 17:32:26.254286 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"af7782f09775c37fb2df7ea471d3f043178c78ecd6178cb2c2bc8c6ee7670877"} Dec 01 17:32:26 crc kubenswrapper[4868]: I1201 17:32:26.254314 4868 scope.go:117] "RemoveContainer" containerID="fd0a3c45f038fc92c88c38c546dd69f6e2fabd876dbe1ef08ab80d88de7be2dd" Dec 01 17:32:26 crc kubenswrapper[4868]: I1201 17:32:26.843759 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" podUID="eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6" containerName="registry" containerID="cri-o://ac6ebb21f802c3122ef3b0fcdf8b34eeea31bf37dcfad221b4818775fda3b20f" gracePeriod=30 Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.268896 4868 generic.go:334] "Generic (PLEG): container finished" podID="eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6" containerID="ac6ebb21f802c3122ef3b0fcdf8b34eeea31bf37dcfad221b4818775fda3b20f" exitCode=0 Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.268986 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" event={"ID":"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6","Type":"ContainerDied","Data":"ac6ebb21f802c3122ef3b0fcdf8b34eeea31bf37dcfad221b4818775fda3b20f"} Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.363845 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.497323 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-registry-tls\") pod \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.497411 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4d64\" (UniqueName: \"kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-kube-api-access-x4d64\") pod \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.497490 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-installation-pull-secrets\") pod \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.497547 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-trusted-ca\") pod \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.498673 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.499031 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.499139 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-ca-trust-extracted\") pod \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.499261 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-bound-sa-token\") pod \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.499283 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-registry-certificates\") pod \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\" (UID: \"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6\") " Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.500112 4868 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.501487 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.512372 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.512364 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.512632 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-kube-api-access-x4d64" (OuterVolumeSpecName: "kube-api-access-x4d64") pod "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6"). InnerVolumeSpecName "kube-api-access-x4d64". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.512663 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.513905 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.524589 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6" (UID: "eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.601498 4868 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.601545 4868 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.601558 4868 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.601575 4868 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.601588 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4d64\" (UniqueName: \"kubernetes.io/projected/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-kube-api-access-x4d64\") on node \"crc\" DevicePath \"\"" Dec 01 17:32:27 crc kubenswrapper[4868]: I1201 17:32:27.601602 4868 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 17:32:28 crc kubenswrapper[4868]: I1201 17:32:28.277147 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" event={"ID":"eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6","Type":"ContainerDied","Data":"ed69eaf9e445234f0e858a52ba50af026ec9ad43f4e70154de73be652d605d02"} Dec 01 17:32:28 crc kubenswrapper[4868]: I1201 17:32:28.277226 4868 scope.go:117] "RemoveContainer" containerID="ac6ebb21f802c3122ef3b0fcdf8b34eeea31bf37dcfad221b4818775fda3b20f" Dec 01 17:32:28 crc kubenswrapper[4868]: I1201 17:32:28.277314 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-x5gxr" Dec 01 17:32:28 crc kubenswrapper[4868]: I1201 17:32:28.307671 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-x5gxr"] Dec 01 17:32:28 crc kubenswrapper[4868]: I1201 17:32:28.308801 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-x5gxr"] Dec 01 17:32:30 crc kubenswrapper[4868]: I1201 17:32:30.179277 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6" path="/var/lib/kubelet/pods/eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6/volumes" Dec 01 17:34:28 crc kubenswrapper[4868]: I1201 17:34:28.385034 4868 scope.go:117] "RemoveContainer" containerID="df09d57f0923446e8127960676005547523e9e1115061159df1ebd582de85c82" Dec 01 17:34:55 crc kubenswrapper[4868]: I1201 17:34:55.905048 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:34:55 crc kubenswrapper[4868]: I1201 17:34:55.905835 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:35:25 crc kubenswrapper[4868]: I1201 17:35:25.905216 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:35:25 crc kubenswrapper[4868]: I1201 17:35:25.906424 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:35:28 crc kubenswrapper[4868]: I1201 17:35:28.429573 4868 scope.go:117] "RemoveContainer" containerID="e63e969ae8ff94477b7a6ceb39d0e7e13188df53cafdc2f44ebea9c8f4a0e176" Dec 01 17:35:28 crc kubenswrapper[4868]: I1201 17:35:28.461227 4868 scope.go:117] "RemoveContainer" containerID="00136348a5d6fe9d51d6e34c1e88a69b9216fb2ac25f3f9950a8b170dee2edb7" Dec 01 17:35:55 crc kubenswrapper[4868]: I1201 17:35:55.904342 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:35:55 crc kubenswrapper[4868]: I1201 17:35:55.904954 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:35:55 crc kubenswrapper[4868]: I1201 17:35:55.905048 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:35:55 crc kubenswrapper[4868]: I1201 17:35:55.905759 4868 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af7782f09775c37fb2df7ea471d3f043178c78ecd6178cb2c2bc8c6ee7670877"} pod="openshift-machine-config-operator/machine-config-daemon-njgpn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 17:35:55 crc kubenswrapper[4868]: I1201 17:35:55.905832 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" containerID="cri-o://af7782f09775c37fb2df7ea471d3f043178c78ecd6178cb2c2bc8c6ee7670877" gracePeriod=600 Dec 01 17:35:56 crc kubenswrapper[4868]: I1201 17:35:56.780936 4868 generic.go:334] "Generic (PLEG): container finished" podID="ec4e08bf-df56-4490-8e79-a4175c727405" containerID="af7782f09775c37fb2df7ea471d3f043178c78ecd6178cb2c2bc8c6ee7670877" exitCode=0 Dec 01 17:35:56 crc kubenswrapper[4868]: I1201 17:35:56.781008 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerDied","Data":"af7782f09775c37fb2df7ea471d3f043178c78ecd6178cb2c2bc8c6ee7670877"} Dec 01 17:35:56 crc kubenswrapper[4868]: I1201 17:35:56.781791 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"5bcb6e04848aace81d63d47c22db36f1a32f6313299199f19533a292326b33e3"} Dec 01 17:35:56 crc kubenswrapper[4868]: I1201 17:35:56.781830 4868 scope.go:117] "RemoveContainer" containerID="5a50633c9f5f1a502388b67f58e02c5f45fd0e84eeadf70b339824d2940a6594" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.062540 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-j2dnr"] Dec 01 17:36:00 crc kubenswrapper[4868]: E1201 17:36:00.063239 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6" containerName="registry" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.063257 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6" containerName="registry" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.063373 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="eefee87f-6ca3-4fcb-8b6f-6b06e06aebe6" containerName="registry" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.063781 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-j2dnr" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.067564 4868 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-t8jrc" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.067762 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.067783 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.076926 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-f6tcg"] Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.078096 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-f6tcg" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.081928 4868 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-8tjhk" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.091898 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-j2dnr"] Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.095704 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-qgmqm"] Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.098542 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-qgmqm" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.100321 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-f6tcg"] Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.101870 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5jc8\" (UniqueName: \"kubernetes.io/projected/958205b1-cd35-41d8-9e21-b30cc80a2489-kube-api-access-g5jc8\") pod \"cert-manager-5b446d88c5-f6tcg\" (UID: \"958205b1-cd35-41d8-9e21-b30cc80a2489\") " pod="cert-manager/cert-manager-5b446d88c5-f6tcg" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.101953 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsqgh\" (UniqueName: \"kubernetes.io/projected/8057b33e-2417-43f9-aab1-7f7db221d051-kube-api-access-fsqgh\") pod \"cert-manager-cainjector-7f985d654d-j2dnr\" (UID: \"8057b33e-2417-43f9-aab1-7f7db221d051\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-j2dnr" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.103309 4868 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-sgsrv" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.123551 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-qgmqm"] Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.203080 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5jc8\" (UniqueName: \"kubernetes.io/projected/958205b1-cd35-41d8-9e21-b30cc80a2489-kube-api-access-g5jc8\") pod \"cert-manager-5b446d88c5-f6tcg\" (UID: \"958205b1-cd35-41d8-9e21-b30cc80a2489\") " pod="cert-manager/cert-manager-5b446d88c5-f6tcg" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.203170 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsqgh\" (UniqueName: \"kubernetes.io/projected/8057b33e-2417-43f9-aab1-7f7db221d051-kube-api-access-fsqgh\") pod \"cert-manager-cainjector-7f985d654d-j2dnr\" (UID: \"8057b33e-2417-43f9-aab1-7f7db221d051\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-j2dnr" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.203211 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4zt2\" (UniqueName: \"kubernetes.io/projected/3858baab-f6b9-42a5-b7ad-2c93a8e98fb6-kube-api-access-l4zt2\") pod \"cert-manager-webhook-5655c58dd6-qgmqm\" (UID: \"3858baab-f6b9-42a5-b7ad-2c93a8e98fb6\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-qgmqm" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.223202 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsqgh\" (UniqueName: \"kubernetes.io/projected/8057b33e-2417-43f9-aab1-7f7db221d051-kube-api-access-fsqgh\") pod \"cert-manager-cainjector-7f985d654d-j2dnr\" (UID: \"8057b33e-2417-43f9-aab1-7f7db221d051\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-j2dnr" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.233856 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5jc8\" (UniqueName: \"kubernetes.io/projected/958205b1-cd35-41d8-9e21-b30cc80a2489-kube-api-access-g5jc8\") pod \"cert-manager-5b446d88c5-f6tcg\" (UID: \"958205b1-cd35-41d8-9e21-b30cc80a2489\") " pod="cert-manager/cert-manager-5b446d88c5-f6tcg" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.310316 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4zt2\" (UniqueName: \"kubernetes.io/projected/3858baab-f6b9-42a5-b7ad-2c93a8e98fb6-kube-api-access-l4zt2\") pod \"cert-manager-webhook-5655c58dd6-qgmqm\" (UID: \"3858baab-f6b9-42a5-b7ad-2c93a8e98fb6\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-qgmqm" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.335600 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4zt2\" (UniqueName: \"kubernetes.io/projected/3858baab-f6b9-42a5-b7ad-2c93a8e98fb6-kube-api-access-l4zt2\") pod \"cert-manager-webhook-5655c58dd6-qgmqm\" (UID: \"3858baab-f6b9-42a5-b7ad-2c93a8e98fb6\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-qgmqm" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.397922 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-j2dnr" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.407347 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-f6tcg" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.419557 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-qgmqm" Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.653240 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-qgmqm"] Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.688792 4868 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.702020 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-j2dnr"] Dec 01 17:36:00 crc kubenswrapper[4868]: W1201 17:36:00.709300 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8057b33e_2417_43f9_aab1_7f7db221d051.slice/crio-ccfd3220f82491df4e26f3f1aee952c07b705df2ed31c0d5aa5b945264b73736 WatchSource:0}: Error finding container ccfd3220f82491df4e26f3f1aee952c07b705df2ed31c0d5aa5b945264b73736: Status 404 returned error can't find the container with id ccfd3220f82491df4e26f3f1aee952c07b705df2ed31c0d5aa5b945264b73736 Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.736688 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-f6tcg"] Dec 01 17:36:00 crc kubenswrapper[4868]: W1201 17:36:00.744492 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod958205b1_cd35_41d8_9e21_b30cc80a2489.slice/crio-c4be9ef45b565583a0a609f43b62a8141dec2f95abeb94d49c78cc6c78057309 WatchSource:0}: Error finding container c4be9ef45b565583a0a609f43b62a8141dec2f95abeb94d49c78cc6c78057309: Status 404 returned error can't find the container with id c4be9ef45b565583a0a609f43b62a8141dec2f95abeb94d49c78cc6c78057309 Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.824060 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-j2dnr" event={"ID":"8057b33e-2417-43f9-aab1-7f7db221d051","Type":"ContainerStarted","Data":"ccfd3220f82491df4e26f3f1aee952c07b705df2ed31c0d5aa5b945264b73736"} Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.826194 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-qgmqm" event={"ID":"3858baab-f6b9-42a5-b7ad-2c93a8e98fb6","Type":"ContainerStarted","Data":"2b9c7fbac1f004deb16bde1266983cfd5c39651f1ecf4e4b5bdced48c10c0387"} Dec 01 17:36:00 crc kubenswrapper[4868]: I1201 17:36:00.828049 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-f6tcg" event={"ID":"958205b1-cd35-41d8-9e21-b30cc80a2489","Type":"ContainerStarted","Data":"c4be9ef45b565583a0a609f43b62a8141dec2f95abeb94d49c78cc6c78057309"} Dec 01 17:36:04 crc kubenswrapper[4868]: I1201 17:36:04.860983 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-j2dnr" event={"ID":"8057b33e-2417-43f9-aab1-7f7db221d051","Type":"ContainerStarted","Data":"41ebd078b3a40546746697424d64fca0289b4a283ed64f77dacce47418e967d5"} Dec 01 17:36:04 crc kubenswrapper[4868]: I1201 17:36:04.867694 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-qgmqm" event={"ID":"3858baab-f6b9-42a5-b7ad-2c93a8e98fb6","Type":"ContainerStarted","Data":"b6ada9b3ff70f65c43883bf19a578ec0164cbc794133fa815015414c42504ab0"} Dec 01 17:36:04 crc kubenswrapper[4868]: I1201 17:36:04.868021 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-qgmqm" Dec 01 17:36:04 crc kubenswrapper[4868]: I1201 17:36:04.869297 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-f6tcg" event={"ID":"958205b1-cd35-41d8-9e21-b30cc80a2489","Type":"ContainerStarted","Data":"6b4a0b130729549700aedfd52acdb85c910774c067da0fbc4c00c3273f3ef094"} Dec 01 17:36:04 crc kubenswrapper[4868]: I1201 17:36:04.881320 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-j2dnr" podStartSLOduration=1.859129523 podStartE2EDuration="4.881292909s" podCreationTimestamp="2025-12-01 17:36:00 +0000 UTC" firstStartedPulling="2025-12-01 17:36:00.712886636 +0000 UTC m=+633.083997047" lastFinishedPulling="2025-12-01 17:36:03.735050022 +0000 UTC m=+636.106160433" observedRunningTime="2025-12-01 17:36:04.877896333 +0000 UTC m=+637.249006744" watchObservedRunningTime="2025-12-01 17:36:04.881292909 +0000 UTC m=+637.252403320" Dec 01 17:36:04 crc kubenswrapper[4868]: I1201 17:36:04.903300 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-qgmqm" podStartSLOduration=1.791691627 podStartE2EDuration="4.903277443s" podCreationTimestamp="2025-12-01 17:36:00 +0000 UTC" firstStartedPulling="2025-12-01 17:36:00.688434372 +0000 UTC m=+633.059544783" lastFinishedPulling="2025-12-01 17:36:03.800020188 +0000 UTC m=+636.171130599" observedRunningTime="2025-12-01 17:36:04.895239699 +0000 UTC m=+637.266350110" watchObservedRunningTime="2025-12-01 17:36:04.903277443 +0000 UTC m=+637.274387854" Dec 01 17:36:04 crc kubenswrapper[4868]: I1201 17:36:04.917040 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-f6tcg" podStartSLOduration=1.9293352339999998 podStartE2EDuration="4.917007446s" podCreationTimestamp="2025-12-01 17:36:00 +0000 UTC" firstStartedPulling="2025-12-01 17:36:00.747319068 +0000 UTC m=+633.118429479" lastFinishedPulling="2025-12-01 17:36:03.73499128 +0000 UTC m=+636.106101691" observedRunningTime="2025-12-01 17:36:04.912299326 +0000 UTC m=+637.283409737" watchObservedRunningTime="2025-12-01 17:36:04.917007446 +0000 UTC m=+637.288117857" Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.425349 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-qgmqm" Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.494189 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-q9z6l"] Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.494721 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovn-controller" containerID="cri-o://bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b" gracePeriod=30 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.494911 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovn-acl-logging" containerID="cri-o://5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4" gracePeriod=30 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.494890 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="kube-rbac-proxy-node" containerID="cri-o://656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df" gracePeriod=30 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.494899 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113" gracePeriod=30 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.495167 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="nbdb" containerID="cri-o://7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854" gracePeriod=30 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.495185 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="northd" containerID="cri-o://df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26" gracePeriod=30 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.495287 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="sbdb" containerID="cri-o://1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0" gracePeriod=30 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.542407 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" containerID="cri-o://ede2d062bb6d3ec0c91e1b108217d9466e8941a233d7e8f2e8c18f033490fb99" gracePeriod=30 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.911648 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9tlw_8ce688b4-36b1-48d4-a6d1-8aec723125c4/kube-multus/2.log" Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.912572 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9tlw_8ce688b4-36b1-48d4-a6d1-8aec723125c4/kube-multus/1.log" Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.912619 4868 generic.go:334] "Generic (PLEG): container finished" podID="8ce688b4-36b1-48d4-a6d1-8aec723125c4" containerID="54e1c435963c9a72f109a2d0e042ba4b27e7e7544588ebfff04e4c7042eb5335" exitCode=2 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.912694 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9tlw" event={"ID":"8ce688b4-36b1-48d4-a6d1-8aec723125c4","Type":"ContainerDied","Data":"54e1c435963c9a72f109a2d0e042ba4b27e7e7544588ebfff04e4c7042eb5335"} Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.912745 4868 scope.go:117] "RemoveContainer" containerID="3d8b1e3f6f1f6f8ed2844b1e533905130d255430f7031699ce02dfa6a3ac16a5" Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.913339 4868 scope.go:117] "RemoveContainer" containerID="54e1c435963c9a72f109a2d0e042ba4b27e7e7544588ebfff04e4c7042eb5335" Dec 01 17:36:10 crc kubenswrapper[4868]: E1201 17:36:10.913683 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-h9tlw_openshift-multus(8ce688b4-36b1-48d4-a6d1-8aec723125c4)\"" pod="openshift-multus/multus-h9tlw" podUID="8ce688b4-36b1-48d4-a6d1-8aec723125c4" Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.917962 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovnkube-controller/3.log" Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.920170 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovn-acl-logging/0.log" Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.920677 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovn-controller/0.log" Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.920999 4868 generic.go:334] "Generic (PLEG): container finished" podID="5baca705-90cc-4198-b641-214a2b915086" containerID="ede2d062bb6d3ec0c91e1b108217d9466e8941a233d7e8f2e8c18f033490fb99" exitCode=0 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.921020 4868 generic.go:334] "Generic (PLEG): container finished" podID="5baca705-90cc-4198-b641-214a2b915086" containerID="1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0" exitCode=0 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.921028 4868 generic.go:334] "Generic (PLEG): container finished" podID="5baca705-90cc-4198-b641-214a2b915086" containerID="7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854" exitCode=0 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.921035 4868 generic.go:334] "Generic (PLEG): container finished" podID="5baca705-90cc-4198-b641-214a2b915086" containerID="df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26" exitCode=0 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.921043 4868 generic.go:334] "Generic (PLEG): container finished" podID="5baca705-90cc-4198-b641-214a2b915086" containerID="f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113" exitCode=0 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.921050 4868 generic.go:334] "Generic (PLEG): container finished" podID="5baca705-90cc-4198-b641-214a2b915086" containerID="656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df" exitCode=0 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.921058 4868 generic.go:334] "Generic (PLEG): container finished" podID="5baca705-90cc-4198-b641-214a2b915086" containerID="5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4" exitCode=143 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.921064 4868 generic.go:334] "Generic (PLEG): container finished" podID="5baca705-90cc-4198-b641-214a2b915086" containerID="bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b" exitCode=143 Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.921063 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerDied","Data":"ede2d062bb6d3ec0c91e1b108217d9466e8941a233d7e8f2e8c18f033490fb99"} Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.921093 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerDied","Data":"1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0"} Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.921103 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerDied","Data":"7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854"} Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.921114 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerDied","Data":"df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26"} Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.921124 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerDied","Data":"f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113"} Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.921133 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerDied","Data":"656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df"} Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.921142 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerDied","Data":"5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4"} Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.921152 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerDied","Data":"bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b"} Dec 01 17:36:10 crc kubenswrapper[4868]: I1201 17:36:10.948144 4868 scope.go:117] "RemoveContainer" containerID="fdd4038187d6e2f96b48ecb68cd129bf8f0c5ba0128bcb5bb5373e534caa699c" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.204402 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovn-acl-logging/0.log" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.205359 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovn-controller/0.log" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.206231 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.272865 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jg6jz"] Dec 01 17:36:11 crc kubenswrapper[4868]: E1201 17:36:11.273181 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="kubecfg-setup" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273197 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="kubecfg-setup" Dec 01 17:36:11 crc kubenswrapper[4868]: E1201 17:36:11.273208 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="kube-rbac-proxy-node" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273214 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="kube-rbac-proxy-node" Dec 01 17:36:11 crc kubenswrapper[4868]: E1201 17:36:11.273231 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273237 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 17:36:11 crc kubenswrapper[4868]: E1201 17:36:11.273243 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273249 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: E1201 17:36:11.273258 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="sbdb" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273264 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="sbdb" Dec 01 17:36:11 crc kubenswrapper[4868]: E1201 17:36:11.273271 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="nbdb" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273276 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="nbdb" Dec 01 17:36:11 crc kubenswrapper[4868]: E1201 17:36:11.273282 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273292 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: E1201 17:36:11.273299 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273305 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: E1201 17:36:11.273311 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273316 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: E1201 17:36:11.273325 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovn-acl-logging" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273331 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovn-acl-logging" Dec 01 17:36:11 crc kubenswrapper[4868]: E1201 17:36:11.273339 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273344 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: E1201 17:36:11.273352 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovn-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273358 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovn-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: E1201 17:36:11.273366 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="northd" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273372 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="northd" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273464 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273475 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273481 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273490 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="northd" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273498 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="nbdb" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273510 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="kube-rbac-proxy-node" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273518 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovn-acl-logging" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273526 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="sbdb" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273535 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovn-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273543 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273709 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.273716 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5baca705-90cc-4198-b641-214a2b915086" containerName="ovnkube-controller" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.275257 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.377777 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-kubelet\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.377825 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-cni-bin\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.377843 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-var-lib-cni-networks-ovn-kubernetes\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.377877 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-systemd-units\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.377903 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-run-ovn-kubernetes\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.377934 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5baca705-90cc-4198-b641-214a2b915086-ovn-node-metrics-cert\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.377963 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-node-log\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378008 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2cpz\" (UniqueName: \"kubernetes.io/projected/5baca705-90cc-4198-b641-214a2b915086-kube-api-access-j2cpz\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378014 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378014 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378026 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-etc-openvswitch\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378050 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378079 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-log-socket\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378081 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378121 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378142 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-var-lib-openvswitch\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378157 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378169 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-openvswitch\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378192 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-node-log" (OuterVolumeSpecName: "node-log") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378219 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-run-netns\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378242 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-ovnkube-config\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378302 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-ovn\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378324 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-env-overrides\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378340 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-systemd\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378363 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-ovnkube-script-lib\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378381 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-slash\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378411 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-cni-netd\") pod \"5baca705-90cc-4198-b641-214a2b915086\" (UID: \"5baca705-90cc-4198-b641-214a2b915086\") " Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378552 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-run-netns\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378584 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-systemd-units\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378605 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-var-lib-openvswitch\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378630 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-cni-bin\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378648 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-ovn-node-metrics-cert\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378677 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-log-socket\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378696 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-ovnkube-config\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378722 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-env-overrides\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378755 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfmg6\" (UniqueName: \"kubernetes.io/projected/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-kube-api-access-qfmg6\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378774 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-run-systemd\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378789 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378808 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-kubelet\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378825 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-slash\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378840 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-cni-netd\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378857 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-run-ovn\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378880 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-etc-openvswitch\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378895 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-ovnkube-script-lib\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378912 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-run-ovn-kubernetes\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378931 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-node-log\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.378987 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-run-openvswitch\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379028 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-log-socket" (OuterVolumeSpecName: "log-socket") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379036 4868 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379059 4868 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379066 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379073 4868 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379096 4868 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379113 4868 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379141 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379164 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379176 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-slash" (OuterVolumeSpecName: "host-slash") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379192 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379232 4868 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379249 4868 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-node-log\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379618 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379632 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.379656 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.380097 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.387966 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5baca705-90cc-4198-b641-214a2b915086-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.388269 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5baca705-90cc-4198-b641-214a2b915086-kube-api-access-j2cpz" (OuterVolumeSpecName: "kube-api-access-j2cpz") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "kube-api-access-j2cpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.399639 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "5baca705-90cc-4198-b641-214a2b915086" (UID: "5baca705-90cc-4198-b641-214a2b915086"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.480128 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-run-openvswitch\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.481102 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-run-netns\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.481223 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-systemd-units\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.481305 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-var-lib-openvswitch\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.481397 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-cni-bin\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.481507 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-ovn-node-metrics-cert\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.481614 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-log-socket\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.481699 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-ovnkube-config\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.481793 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-env-overrides\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.481863 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-run-netns\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.481890 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-log-socket\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.481886 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-systemd-units\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.481823 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-var-lib-openvswitch\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.480498 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-run-openvswitch\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.482027 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-cni-bin\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.482200 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfmg6\" (UniqueName: \"kubernetes.io/projected/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-kube-api-access-qfmg6\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.482309 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-run-systemd\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.482402 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.482491 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-kubelet\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.482571 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-slash\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.482661 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-cni-netd\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.482756 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-run-ovn\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.482872 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-etc-openvswitch\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.482990 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-env-overrides\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483000 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-ovnkube-script-lib\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483080 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-run-ovn-kubernetes\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483115 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-node-log\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483217 4868 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483233 4868 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483245 4868 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483255 4868 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483264 4868 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483266 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-ovnkube-config\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483273 4868 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5baca705-90cc-4198-b641-214a2b915086-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483303 4868 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-slash\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483315 4868 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483329 4868 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5baca705-90cc-4198-b641-214a2b915086-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483343 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2cpz\" (UniqueName: \"kubernetes.io/projected/5baca705-90cc-4198-b641-214a2b915086-kube-api-access-j2cpz\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483355 4868 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-log-socket\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483370 4868 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483386 4868 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5baca705-90cc-4198-b641-214a2b915086-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483054 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-kubelet\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.483883 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-node-log\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.484015 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-run-ovn-kubernetes\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.484080 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-run-systemd\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.484211 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-cni-netd\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.484248 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-slash\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.484250 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.484272 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-run-ovn\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.484284 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-etc-openvswitch\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.484935 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-ovnkube-script-lib\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.487431 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-ovn-node-metrics-cert\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.504891 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfmg6\" (UniqueName: \"kubernetes.io/projected/f119f5fd-d629-4e7e-b6af-8d28ea8fab88-kube-api-access-qfmg6\") pod \"ovnkube-node-jg6jz\" (UID: \"f119f5fd-d629-4e7e-b6af-8d28ea8fab88\") " pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.604299 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.930361 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" event={"ID":"f119f5fd-d629-4e7e-b6af-8d28ea8fab88","Type":"ContainerStarted","Data":"fd56f380b6011a1e50b643e547317291565a2a2af929c01ca6f714fe69ed91aa"} Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.933038 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9tlw_8ce688b4-36b1-48d4-a6d1-8aec723125c4/kube-multus/2.log" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.939394 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovn-acl-logging/0.log" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.940802 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-q9z6l_5baca705-90cc-4198-b641-214a2b915086/ovn-controller/0.log" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.941509 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" event={"ID":"5baca705-90cc-4198-b641-214a2b915086","Type":"ContainerDied","Data":"ef85c890c89cf0cd7e1efcda5f1342842e1535dfceed118f62c4c9af9c11bb55"} Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.941590 4868 scope.go:117] "RemoveContainer" containerID="ede2d062bb6d3ec0c91e1b108217d9466e8941a233d7e8f2e8c18f033490fb99" Dec 01 17:36:11 crc kubenswrapper[4868]: I1201 17:36:11.941712 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-q9z6l" Dec 01 17:36:12 crc kubenswrapper[4868]: I1201 17:36:12.020085 4868 scope.go:117] "RemoveContainer" containerID="1e9e70fa5bf89a485428ae0ab1f3833ae6b69b7c6a7d4e68c07d4591915acbc0" Dec 01 17:36:12 crc kubenswrapper[4868]: I1201 17:36:12.038996 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-q9z6l"] Dec 01 17:36:12 crc kubenswrapper[4868]: I1201 17:36:12.042493 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-q9z6l"] Dec 01 17:36:12 crc kubenswrapper[4868]: I1201 17:36:12.044397 4868 scope.go:117] "RemoveContainer" containerID="7457eaab6ba97098079424db6084f7c7518de8976a8de40401624aecbc6a4854" Dec 01 17:36:12 crc kubenswrapper[4868]: I1201 17:36:12.061653 4868 scope.go:117] "RemoveContainer" containerID="df9528675fa8fda4fb42f412987c6f3068355db7ceac315f7759dc1bc6b07a26" Dec 01 17:36:12 crc kubenswrapper[4868]: I1201 17:36:12.079634 4868 scope.go:117] "RemoveContainer" containerID="f5f1f5698a2a85d64707203ee4903bd9bc5b3c77997fab1eefc208142303b113" Dec 01 17:36:12 crc kubenswrapper[4868]: I1201 17:36:12.162838 4868 scope.go:117] "RemoveContainer" containerID="656506d293021498f08136bfa115e4b05980a8b2b0ac4b478c6072f786d6f8df" Dec 01 17:36:12 crc kubenswrapper[4868]: I1201 17:36:12.182516 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5baca705-90cc-4198-b641-214a2b915086" path="/var/lib/kubelet/pods/5baca705-90cc-4198-b641-214a2b915086/volumes" Dec 01 17:36:12 crc kubenswrapper[4868]: I1201 17:36:12.186853 4868 scope.go:117] "RemoveContainer" containerID="5422667cdf3825c87ad2dfcc6c978942928c64946ef722e064780a11bb96c4b4" Dec 01 17:36:12 crc kubenswrapper[4868]: I1201 17:36:12.203386 4868 scope.go:117] "RemoveContainer" containerID="bff59c178a6d7279091288badac18b17e1f25150d9aa7df8f56b74c1e5a5484b" Dec 01 17:36:12 crc kubenswrapper[4868]: I1201 17:36:12.219792 4868 scope.go:117] "RemoveContainer" containerID="e2afffbb2bce347630f02a9e2a50741057356e8d805dcd7ae1d05549f21b11f5" Dec 01 17:36:12 crc kubenswrapper[4868]: I1201 17:36:12.951902 4868 generic.go:334] "Generic (PLEG): container finished" podID="f119f5fd-d629-4e7e-b6af-8d28ea8fab88" containerID="1904bd23ac01bb0a853ecca8d8b81d403b35751183af39f37244e69c8a48a4b1" exitCode=0 Dec 01 17:36:12 crc kubenswrapper[4868]: I1201 17:36:12.952018 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" event={"ID":"f119f5fd-d629-4e7e-b6af-8d28ea8fab88","Type":"ContainerDied","Data":"1904bd23ac01bb0a853ecca8d8b81d403b35751183af39f37244e69c8a48a4b1"} Dec 01 17:36:13 crc kubenswrapper[4868]: I1201 17:36:13.973683 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" event={"ID":"f119f5fd-d629-4e7e-b6af-8d28ea8fab88","Type":"ContainerStarted","Data":"a88415926a49c1b1ff3b733d89753fc44ed69b581395cbdd427a2b48e33d5c65"} Dec 01 17:36:13 crc kubenswrapper[4868]: I1201 17:36:13.974722 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" event={"ID":"f119f5fd-d629-4e7e-b6af-8d28ea8fab88","Type":"ContainerStarted","Data":"c601b0fdca1d1d5f11f872af8f9e25d8ad9b8e6a2f141382945a106b69f5a7e2"} Dec 01 17:36:13 crc kubenswrapper[4868]: I1201 17:36:13.974752 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" event={"ID":"f119f5fd-d629-4e7e-b6af-8d28ea8fab88","Type":"ContainerStarted","Data":"4599cbc4be941ec252f3b23b791835a2ec10b8b4c544e9ebc5c57f0b2a1a7230"} Dec 01 17:36:13 crc kubenswrapper[4868]: I1201 17:36:13.974781 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" event={"ID":"f119f5fd-d629-4e7e-b6af-8d28ea8fab88","Type":"ContainerStarted","Data":"dadf84e503657849b2c258ac2835d98e30c0db42745de43984789144cca7dcfb"} Dec 01 17:36:13 crc kubenswrapper[4868]: I1201 17:36:13.974805 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" event={"ID":"f119f5fd-d629-4e7e-b6af-8d28ea8fab88","Type":"ContainerStarted","Data":"e631c079bea149e0b63173dbcb4ec6bd1b87b0b749fb9adf2eae695b00a588ad"} Dec 01 17:36:13 crc kubenswrapper[4868]: I1201 17:36:13.974829 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" event={"ID":"f119f5fd-d629-4e7e-b6af-8d28ea8fab88","Type":"ContainerStarted","Data":"d97dba67bafbd3fd350f1686a2e05e03d59bac6a7a3ffb6e3dafffa4c0ea8587"} Dec 01 17:36:17 crc kubenswrapper[4868]: I1201 17:36:17.005637 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" event={"ID":"f119f5fd-d629-4e7e-b6af-8d28ea8fab88","Type":"ContainerStarted","Data":"74de8388411467ee1d76b1480e9f081bf3cc29bed1deba0a4921ac3629586155"} Dec 01 17:36:19 crc kubenswrapper[4868]: I1201 17:36:19.028782 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" event={"ID":"f119f5fd-d629-4e7e-b6af-8d28ea8fab88","Type":"ContainerStarted","Data":"3aa268ab2f8f9279dab5af4179897f7a00e08c05ebbdd7a2ef4c5b361adb755c"} Dec 01 17:36:19 crc kubenswrapper[4868]: I1201 17:36:19.029561 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:19 crc kubenswrapper[4868]: I1201 17:36:19.029675 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:19 crc kubenswrapper[4868]: I1201 17:36:19.029745 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:19 crc kubenswrapper[4868]: I1201 17:36:19.060815 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:19 crc kubenswrapper[4868]: I1201 17:36:19.061785 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:19 crc kubenswrapper[4868]: I1201 17:36:19.078163 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" podStartSLOduration=8.078139075 podStartE2EDuration="8.078139075s" podCreationTimestamp="2025-12-01 17:36:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:36:19.072912549 +0000 UTC m=+651.444022990" watchObservedRunningTime="2025-12-01 17:36:19.078139075 +0000 UTC m=+651.449249486" Dec 01 17:36:23 crc kubenswrapper[4868]: I1201 17:36:23.171683 4868 scope.go:117] "RemoveContainer" containerID="54e1c435963c9a72f109a2d0e042ba4b27e7e7544588ebfff04e4c7042eb5335" Dec 01 17:36:23 crc kubenswrapper[4868]: E1201 17:36:23.172875 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-h9tlw_openshift-multus(8ce688b4-36b1-48d4-a6d1-8aec723125c4)\"" pod="openshift-multus/multus-h9tlw" podUID="8ce688b4-36b1-48d4-a6d1-8aec723125c4" Dec 01 17:36:37 crc kubenswrapper[4868]: I1201 17:36:37.173590 4868 scope.go:117] "RemoveContainer" containerID="54e1c435963c9a72f109a2d0e042ba4b27e7e7544588ebfff04e4c7042eb5335" Dec 01 17:36:38 crc kubenswrapper[4868]: I1201 17:36:38.166693 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9tlw_8ce688b4-36b1-48d4-a6d1-8aec723125c4/kube-multus/2.log" Dec 01 17:36:38 crc kubenswrapper[4868]: I1201 17:36:38.167447 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9tlw" event={"ID":"8ce688b4-36b1-48d4-a6d1-8aec723125c4","Type":"ContainerStarted","Data":"de8033bb3ee1790578b2562acce3e44112f553b59a89eb8337a4326cb4a7ee12"} Dec 01 17:36:41 crc kubenswrapper[4868]: I1201 17:36:41.639290 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jg6jz" Dec 01 17:36:50 crc kubenswrapper[4868]: I1201 17:36:50.860807 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6"] Dec 01 17:36:50 crc kubenswrapper[4868]: I1201 17:36:50.863092 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" Dec 01 17:36:50 crc kubenswrapper[4868]: I1201 17:36:50.865441 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 17:36:50 crc kubenswrapper[4868]: I1201 17:36:50.873498 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6"] Dec 01 17:36:50 crc kubenswrapper[4868]: I1201 17:36:50.931095 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzfb8\" (UniqueName: \"kubernetes.io/projected/e1d4847d-1891-4e4c-9d68-04181da92d62-kube-api-access-zzfb8\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6\" (UID: \"e1d4847d-1891-4e4c-9d68-04181da92d62\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" Dec 01 17:36:50 crc kubenswrapper[4868]: I1201 17:36:50.931171 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1d4847d-1891-4e4c-9d68-04181da92d62-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6\" (UID: \"e1d4847d-1891-4e4c-9d68-04181da92d62\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" Dec 01 17:36:50 crc kubenswrapper[4868]: I1201 17:36:50.931232 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1d4847d-1891-4e4c-9d68-04181da92d62-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6\" (UID: \"e1d4847d-1891-4e4c-9d68-04181da92d62\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" Dec 01 17:36:51 crc kubenswrapper[4868]: I1201 17:36:51.032540 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzfb8\" (UniqueName: \"kubernetes.io/projected/e1d4847d-1891-4e4c-9d68-04181da92d62-kube-api-access-zzfb8\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6\" (UID: \"e1d4847d-1891-4e4c-9d68-04181da92d62\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" Dec 01 17:36:51 crc kubenswrapper[4868]: I1201 17:36:51.032607 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1d4847d-1891-4e4c-9d68-04181da92d62-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6\" (UID: \"e1d4847d-1891-4e4c-9d68-04181da92d62\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" Dec 01 17:36:51 crc kubenswrapper[4868]: I1201 17:36:51.032628 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1d4847d-1891-4e4c-9d68-04181da92d62-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6\" (UID: \"e1d4847d-1891-4e4c-9d68-04181da92d62\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" Dec 01 17:36:51 crc kubenswrapper[4868]: I1201 17:36:51.033211 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1d4847d-1891-4e4c-9d68-04181da92d62-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6\" (UID: \"e1d4847d-1891-4e4c-9d68-04181da92d62\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" Dec 01 17:36:51 crc kubenswrapper[4868]: I1201 17:36:51.033502 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1d4847d-1891-4e4c-9d68-04181da92d62-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6\" (UID: \"e1d4847d-1891-4e4c-9d68-04181da92d62\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" Dec 01 17:36:51 crc kubenswrapper[4868]: I1201 17:36:51.054993 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzfb8\" (UniqueName: \"kubernetes.io/projected/e1d4847d-1891-4e4c-9d68-04181da92d62-kube-api-access-zzfb8\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6\" (UID: \"e1d4847d-1891-4e4c-9d68-04181da92d62\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" Dec 01 17:36:51 crc kubenswrapper[4868]: I1201 17:36:51.184005 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" Dec 01 17:36:51 crc kubenswrapper[4868]: I1201 17:36:51.463117 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6"] Dec 01 17:36:52 crc kubenswrapper[4868]: I1201 17:36:52.262447 4868 generic.go:334] "Generic (PLEG): container finished" podID="e1d4847d-1891-4e4c-9d68-04181da92d62" containerID="4ce2f8aebcfb1b3f806ad803ec1edc916f6dae6555807b1f5f8d666dbdc53dce" exitCode=0 Dec 01 17:36:52 crc kubenswrapper[4868]: I1201 17:36:52.262700 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" event={"ID":"e1d4847d-1891-4e4c-9d68-04181da92d62","Type":"ContainerDied","Data":"4ce2f8aebcfb1b3f806ad803ec1edc916f6dae6555807b1f5f8d666dbdc53dce"} Dec 01 17:36:52 crc kubenswrapper[4868]: I1201 17:36:52.262989 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" event={"ID":"e1d4847d-1891-4e4c-9d68-04181da92d62","Type":"ContainerStarted","Data":"17078ea7a8dce1195a616aec159ef33dcf9e46b557a74ab35fe5fd4cfa1081ae"} Dec 01 17:36:55 crc kubenswrapper[4868]: I1201 17:36:55.285295 4868 generic.go:334] "Generic (PLEG): container finished" podID="e1d4847d-1891-4e4c-9d68-04181da92d62" containerID="186b17f3c3c6d72dff8b8bb1fa2d1cedac65cfae10d7187e35663d678725f88e" exitCode=0 Dec 01 17:36:55 crc kubenswrapper[4868]: I1201 17:36:55.285400 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" event={"ID":"e1d4847d-1891-4e4c-9d68-04181da92d62","Type":"ContainerDied","Data":"186b17f3c3c6d72dff8b8bb1fa2d1cedac65cfae10d7187e35663d678725f88e"} Dec 01 17:36:56 crc kubenswrapper[4868]: I1201 17:36:56.296235 4868 generic.go:334] "Generic (PLEG): container finished" podID="e1d4847d-1891-4e4c-9d68-04181da92d62" containerID="416cdcd27001278103c86bb58346e7f39cd5ac247e5fefbf907fd91d9d173fae" exitCode=0 Dec 01 17:36:56 crc kubenswrapper[4868]: I1201 17:36:56.296338 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" event={"ID":"e1d4847d-1891-4e4c-9d68-04181da92d62","Type":"ContainerDied","Data":"416cdcd27001278103c86bb58346e7f39cd5ac247e5fefbf907fd91d9d173fae"} Dec 01 17:36:57 crc kubenswrapper[4868]: I1201 17:36:57.591823 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" Dec 01 17:36:57 crc kubenswrapper[4868]: I1201 17:36:57.748101 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1d4847d-1891-4e4c-9d68-04181da92d62-util\") pod \"e1d4847d-1891-4e4c-9d68-04181da92d62\" (UID: \"e1d4847d-1891-4e4c-9d68-04181da92d62\") " Dec 01 17:36:57 crc kubenswrapper[4868]: I1201 17:36:57.748231 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1d4847d-1891-4e4c-9d68-04181da92d62-bundle\") pod \"e1d4847d-1891-4e4c-9d68-04181da92d62\" (UID: \"e1d4847d-1891-4e4c-9d68-04181da92d62\") " Dec 01 17:36:57 crc kubenswrapper[4868]: I1201 17:36:57.748334 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzfb8\" (UniqueName: \"kubernetes.io/projected/e1d4847d-1891-4e4c-9d68-04181da92d62-kube-api-access-zzfb8\") pod \"e1d4847d-1891-4e4c-9d68-04181da92d62\" (UID: \"e1d4847d-1891-4e4c-9d68-04181da92d62\") " Dec 01 17:36:57 crc kubenswrapper[4868]: I1201 17:36:57.750506 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1d4847d-1891-4e4c-9d68-04181da92d62-bundle" (OuterVolumeSpecName: "bundle") pod "e1d4847d-1891-4e4c-9d68-04181da92d62" (UID: "e1d4847d-1891-4e4c-9d68-04181da92d62"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:36:57 crc kubenswrapper[4868]: I1201 17:36:57.758969 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1d4847d-1891-4e4c-9d68-04181da92d62-kube-api-access-zzfb8" (OuterVolumeSpecName: "kube-api-access-zzfb8") pod "e1d4847d-1891-4e4c-9d68-04181da92d62" (UID: "e1d4847d-1891-4e4c-9d68-04181da92d62"). InnerVolumeSpecName "kube-api-access-zzfb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:36:57 crc kubenswrapper[4868]: I1201 17:36:57.769541 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1d4847d-1891-4e4c-9d68-04181da92d62-util" (OuterVolumeSpecName: "util") pod "e1d4847d-1891-4e4c-9d68-04181da92d62" (UID: "e1d4847d-1891-4e4c-9d68-04181da92d62"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:36:57 crc kubenswrapper[4868]: I1201 17:36:57.850101 4868 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1d4847d-1891-4e4c-9d68-04181da92d62-util\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:57 crc kubenswrapper[4868]: I1201 17:36:57.850175 4868 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1d4847d-1891-4e4c-9d68-04181da92d62-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:57 crc kubenswrapper[4868]: I1201 17:36:57.850195 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzfb8\" (UniqueName: \"kubernetes.io/projected/e1d4847d-1891-4e4c-9d68-04181da92d62-kube-api-access-zzfb8\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:58 crc kubenswrapper[4868]: I1201 17:36:58.314779 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" event={"ID":"e1d4847d-1891-4e4c-9d68-04181da92d62","Type":"ContainerDied","Data":"17078ea7a8dce1195a616aec159ef33dcf9e46b557a74ab35fe5fd4cfa1081ae"} Dec 01 17:36:58 crc kubenswrapper[4868]: I1201 17:36:58.314838 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6" Dec 01 17:36:58 crc kubenswrapper[4868]: I1201 17:36:58.314840 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17078ea7a8dce1195a616aec159ef33dcf9e46b557a74ab35fe5fd4cfa1081ae" Dec 01 17:36:59 crc kubenswrapper[4868]: I1201 17:36:59.526097 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-rb8wl"] Dec 01 17:36:59 crc kubenswrapper[4868]: E1201 17:36:59.526360 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d4847d-1891-4e4c-9d68-04181da92d62" containerName="extract" Dec 01 17:36:59 crc kubenswrapper[4868]: I1201 17:36:59.526374 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d4847d-1891-4e4c-9d68-04181da92d62" containerName="extract" Dec 01 17:36:59 crc kubenswrapper[4868]: E1201 17:36:59.526384 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d4847d-1891-4e4c-9d68-04181da92d62" containerName="pull" Dec 01 17:36:59 crc kubenswrapper[4868]: I1201 17:36:59.526389 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d4847d-1891-4e4c-9d68-04181da92d62" containerName="pull" Dec 01 17:36:59 crc kubenswrapper[4868]: E1201 17:36:59.526406 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d4847d-1891-4e4c-9d68-04181da92d62" containerName="util" Dec 01 17:36:59 crc kubenswrapper[4868]: I1201 17:36:59.526414 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d4847d-1891-4e4c-9d68-04181da92d62" containerName="util" Dec 01 17:36:59 crc kubenswrapper[4868]: I1201 17:36:59.526525 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d4847d-1891-4e4c-9d68-04181da92d62" containerName="extract" Dec 01 17:36:59 crc kubenswrapper[4868]: I1201 17:36:59.526989 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-rb8wl" Dec 01 17:36:59 crc kubenswrapper[4868]: I1201 17:36:59.529047 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 01 17:36:59 crc kubenswrapper[4868]: I1201 17:36:59.529528 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 01 17:36:59 crc kubenswrapper[4868]: I1201 17:36:59.530678 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-t6fg9" Dec 01 17:36:59 crc kubenswrapper[4868]: I1201 17:36:59.539972 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-rb8wl"] Dec 01 17:36:59 crc kubenswrapper[4868]: I1201 17:36:59.683228 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p762r\" (UniqueName: \"kubernetes.io/projected/13c419e8-4943-4ee5-8d1c-5dd5051c81db-kube-api-access-p762r\") pod \"nmstate-operator-5b5b58f5c8-rb8wl\" (UID: \"13c419e8-4943-4ee5-8d1c-5dd5051c81db\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-rb8wl" Dec 01 17:36:59 crc kubenswrapper[4868]: I1201 17:36:59.784609 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p762r\" (UniqueName: \"kubernetes.io/projected/13c419e8-4943-4ee5-8d1c-5dd5051c81db-kube-api-access-p762r\") pod \"nmstate-operator-5b5b58f5c8-rb8wl\" (UID: \"13c419e8-4943-4ee5-8d1c-5dd5051c81db\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-rb8wl" Dec 01 17:36:59 crc kubenswrapper[4868]: I1201 17:36:59.806888 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p762r\" (UniqueName: \"kubernetes.io/projected/13c419e8-4943-4ee5-8d1c-5dd5051c81db-kube-api-access-p762r\") pod \"nmstate-operator-5b5b58f5c8-rb8wl\" (UID: \"13c419e8-4943-4ee5-8d1c-5dd5051c81db\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-rb8wl" Dec 01 17:36:59 crc kubenswrapper[4868]: I1201 17:36:59.843869 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-rb8wl" Dec 01 17:37:00 crc kubenswrapper[4868]: I1201 17:37:00.080962 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-rb8wl"] Dec 01 17:37:00 crc kubenswrapper[4868]: I1201 17:37:00.329642 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-rb8wl" event={"ID":"13c419e8-4943-4ee5-8d1c-5dd5051c81db","Type":"ContainerStarted","Data":"f81fb8d3b61e47b594aba7e1c43533a70ba8f81bfd2d80001a9346336662a776"} Dec 01 17:37:04 crc kubenswrapper[4868]: I1201 17:37:04.360382 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-rb8wl" event={"ID":"13c419e8-4943-4ee5-8d1c-5dd5051c81db","Type":"ContainerStarted","Data":"1adb2269aabc358222ee216c8032160d1158afc5033369c3f73d9c62ef25f1b2"} Dec 01 17:37:04 crc kubenswrapper[4868]: I1201 17:37:04.388805 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-rb8wl" podStartSLOduration=2.170800613 podStartE2EDuration="5.388768514s" podCreationTimestamp="2025-12-01 17:36:59 +0000 UTC" firstStartedPulling="2025-12-01 17:37:00.091486426 +0000 UTC m=+692.462596837" lastFinishedPulling="2025-12-01 17:37:03.309454327 +0000 UTC m=+695.680564738" observedRunningTime="2025-12-01 17:37:04.384281298 +0000 UTC m=+696.755391739" watchObservedRunningTime="2025-12-01 17:37:04.388768514 +0000 UTC m=+696.759878935" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.396840 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-j7lcc"] Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.398473 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-j7lcc" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.402305 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-zthff" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.405464 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw"] Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.406793 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.407875 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-j7lcc"] Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.409685 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.421931 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-67rpb"] Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.423091 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.447579 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw"] Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.477131 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfkbj\" (UniqueName: \"kubernetes.io/projected/d04c81f8-7543-44b9-abc9-b0c95346cd56-kube-api-access-wfkbj\") pod \"nmstate-handler-67rpb\" (UID: \"d04c81f8-7543-44b9-abc9-b0c95346cd56\") " pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.477205 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d04c81f8-7543-44b9-abc9-b0c95346cd56-dbus-socket\") pod \"nmstate-handler-67rpb\" (UID: \"d04c81f8-7543-44b9-abc9-b0c95346cd56\") " pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.477268 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2bh5\" (UniqueName: \"kubernetes.io/projected/f1784373-a1b0-4ca1-8c08-cf01fbc83b72-kube-api-access-n2bh5\") pod \"nmstate-webhook-5f6d4c5ccb-kjwkw\" (UID: \"f1784373-a1b0-4ca1-8c08-cf01fbc83b72\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.477305 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfmbh\" (UniqueName: \"kubernetes.io/projected/ed23de21-61e1-463e-9a8e-5a8ba51cc727-kube-api-access-nfmbh\") pod \"nmstate-metrics-7f946cbc9-j7lcc\" (UID: \"ed23de21-61e1-463e-9a8e-5a8ba51cc727\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-j7lcc" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.477376 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f1784373-a1b0-4ca1-8c08-cf01fbc83b72-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-kjwkw\" (UID: \"f1784373-a1b0-4ca1-8c08-cf01fbc83b72\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.477422 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d04c81f8-7543-44b9-abc9-b0c95346cd56-nmstate-lock\") pod \"nmstate-handler-67rpb\" (UID: \"d04c81f8-7543-44b9-abc9-b0c95346cd56\") " pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.477467 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d04c81f8-7543-44b9-abc9-b0c95346cd56-ovs-socket\") pod \"nmstate-handler-67rpb\" (UID: \"d04c81f8-7543-44b9-abc9-b0c95346cd56\") " pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.553664 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b"] Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.554525 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.558021 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-8kz4f" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.558210 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.558314 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.565440 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b"] Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.581567 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfmbh\" (UniqueName: \"kubernetes.io/projected/ed23de21-61e1-463e-9a8e-5a8ba51cc727-kube-api-access-nfmbh\") pod \"nmstate-metrics-7f946cbc9-j7lcc\" (UID: \"ed23de21-61e1-463e-9a8e-5a8ba51cc727\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-j7lcc" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.581625 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa121055-74c0-44aa-8401-3a8476c7db4c-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-xsj4b\" (UID: \"fa121055-74c0-44aa-8401-3a8476c7db4c\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.581665 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f1784373-a1b0-4ca1-8c08-cf01fbc83b72-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-kjwkw\" (UID: \"f1784373-a1b0-4ca1-8c08-cf01fbc83b72\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.581718 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d04c81f8-7543-44b9-abc9-b0c95346cd56-nmstate-lock\") pod \"nmstate-handler-67rpb\" (UID: \"d04c81f8-7543-44b9-abc9-b0c95346cd56\") " pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.581759 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d04c81f8-7543-44b9-abc9-b0c95346cd56-ovs-socket\") pod \"nmstate-handler-67rpb\" (UID: \"d04c81f8-7543-44b9-abc9-b0c95346cd56\") " pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.581781 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wt22\" (UniqueName: \"kubernetes.io/projected/fa121055-74c0-44aa-8401-3a8476c7db4c-kube-api-access-2wt22\") pod \"nmstate-console-plugin-7fbb5f6569-xsj4b\" (UID: \"fa121055-74c0-44aa-8401-3a8476c7db4c\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.581814 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfkbj\" (UniqueName: \"kubernetes.io/projected/d04c81f8-7543-44b9-abc9-b0c95346cd56-kube-api-access-wfkbj\") pod \"nmstate-handler-67rpb\" (UID: \"d04c81f8-7543-44b9-abc9-b0c95346cd56\") " pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.581840 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d04c81f8-7543-44b9-abc9-b0c95346cd56-dbus-socket\") pod \"nmstate-handler-67rpb\" (UID: \"d04c81f8-7543-44b9-abc9-b0c95346cd56\") " pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.581860 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fa121055-74c0-44aa-8401-3a8476c7db4c-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-xsj4b\" (UID: \"fa121055-74c0-44aa-8401-3a8476c7db4c\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.581891 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2bh5\" (UniqueName: \"kubernetes.io/projected/f1784373-a1b0-4ca1-8c08-cf01fbc83b72-kube-api-access-n2bh5\") pod \"nmstate-webhook-5f6d4c5ccb-kjwkw\" (UID: \"f1784373-a1b0-4ca1-8c08-cf01fbc83b72\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw" Dec 01 17:37:05 crc kubenswrapper[4868]: E1201 17:37:05.582447 4868 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 01 17:37:05 crc kubenswrapper[4868]: E1201 17:37:05.582501 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f1784373-a1b0-4ca1-8c08-cf01fbc83b72-tls-key-pair podName:f1784373-a1b0-4ca1-8c08-cf01fbc83b72 nodeName:}" failed. No retries permitted until 2025-12-01 17:37:06.082484434 +0000 UTC m=+698.453594845 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/f1784373-a1b0-4ca1-8c08-cf01fbc83b72-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-kjwkw" (UID: "f1784373-a1b0-4ca1-8c08-cf01fbc83b72") : secret "openshift-nmstate-webhook" not found Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.582649 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d04c81f8-7543-44b9-abc9-b0c95346cd56-nmstate-lock\") pod \"nmstate-handler-67rpb\" (UID: \"d04c81f8-7543-44b9-abc9-b0c95346cd56\") " pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.582682 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d04c81f8-7543-44b9-abc9-b0c95346cd56-ovs-socket\") pod \"nmstate-handler-67rpb\" (UID: \"d04c81f8-7543-44b9-abc9-b0c95346cd56\") " pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.583112 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d04c81f8-7543-44b9-abc9-b0c95346cd56-dbus-socket\") pod \"nmstate-handler-67rpb\" (UID: \"d04c81f8-7543-44b9-abc9-b0c95346cd56\") " pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.613645 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfmbh\" (UniqueName: \"kubernetes.io/projected/ed23de21-61e1-463e-9a8e-5a8ba51cc727-kube-api-access-nfmbh\") pod \"nmstate-metrics-7f946cbc9-j7lcc\" (UID: \"ed23de21-61e1-463e-9a8e-5a8ba51cc727\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-j7lcc" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.619858 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2bh5\" (UniqueName: \"kubernetes.io/projected/f1784373-a1b0-4ca1-8c08-cf01fbc83b72-kube-api-access-n2bh5\") pod \"nmstate-webhook-5f6d4c5ccb-kjwkw\" (UID: \"f1784373-a1b0-4ca1-8c08-cf01fbc83b72\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.626305 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfkbj\" (UniqueName: \"kubernetes.io/projected/d04c81f8-7543-44b9-abc9-b0c95346cd56-kube-api-access-wfkbj\") pod \"nmstate-handler-67rpb\" (UID: \"d04c81f8-7543-44b9-abc9-b0c95346cd56\") " pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.683707 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wt22\" (UniqueName: \"kubernetes.io/projected/fa121055-74c0-44aa-8401-3a8476c7db4c-kube-api-access-2wt22\") pod \"nmstate-console-plugin-7fbb5f6569-xsj4b\" (UID: \"fa121055-74c0-44aa-8401-3a8476c7db4c\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.683761 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fa121055-74c0-44aa-8401-3a8476c7db4c-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-xsj4b\" (UID: \"fa121055-74c0-44aa-8401-3a8476c7db4c\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.683812 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa121055-74c0-44aa-8401-3a8476c7db4c-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-xsj4b\" (UID: \"fa121055-74c0-44aa-8401-3a8476c7db4c\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b" Dec 01 17:37:05 crc kubenswrapper[4868]: E1201 17:37:05.683934 4868 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 01 17:37:05 crc kubenswrapper[4868]: E1201 17:37:05.684006 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fa121055-74c0-44aa-8401-3a8476c7db4c-plugin-serving-cert podName:fa121055-74c0-44aa-8401-3a8476c7db4c nodeName:}" failed. No retries permitted until 2025-12-01 17:37:06.183990314 +0000 UTC m=+698.555100725 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/fa121055-74c0-44aa-8401-3a8476c7db4c-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-xsj4b" (UID: "fa121055-74c0-44aa-8401-3a8476c7db4c") : secret "plugin-serving-cert" not found Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.686005 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fa121055-74c0-44aa-8401-3a8476c7db4c-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-xsj4b\" (UID: \"fa121055-74c0-44aa-8401-3a8476c7db4c\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.722617 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wt22\" (UniqueName: \"kubernetes.io/projected/fa121055-74c0-44aa-8401-3a8476c7db4c-kube-api-access-2wt22\") pod \"nmstate-console-plugin-7fbb5f6569-xsj4b\" (UID: \"fa121055-74c0-44aa-8401-3a8476c7db4c\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.732909 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-j7lcc" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.775126 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.831851 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-594dc658c5-85bst"] Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.834251 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.858018 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-594dc658c5-85bst"] Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.890619 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-console-serving-cert\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.890930 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f298b\" (UniqueName: \"kubernetes.io/projected/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-kube-api-access-f298b\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.890987 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-console-oauth-config\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.891018 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-service-ca\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.891057 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-console-config\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.891074 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-oauth-serving-cert\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.891113 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-trusted-ca-bundle\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.992343 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-trusted-ca-bundle\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.993240 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-console-serving-cert\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.993340 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f298b\" (UniqueName: \"kubernetes.io/projected/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-kube-api-access-f298b\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.993386 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-console-oauth-config\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.993445 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-service-ca\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.993502 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-console-config\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.993527 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-oauth-serving-cert\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.994317 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-trusted-ca-bundle\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.994815 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-oauth-serving-cert\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.995181 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-service-ca\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:05 crc kubenswrapper[4868]: I1201 17:37:05.995221 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-console-config\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.000933 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-console-serving-cert\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.001068 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-console-oauth-config\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.012261 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f298b\" (UniqueName: \"kubernetes.io/projected/55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58-kube-api-access-f298b\") pod \"console-594dc658c5-85bst\" (UID: \"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58\") " pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.017894 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-j7lcc"] Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.095382 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f1784373-a1b0-4ca1-8c08-cf01fbc83b72-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-kjwkw\" (UID: \"f1784373-a1b0-4ca1-8c08-cf01fbc83b72\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw" Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.099621 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f1784373-a1b0-4ca1-8c08-cf01fbc83b72-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-kjwkw\" (UID: \"f1784373-a1b0-4ca1-8c08-cf01fbc83b72\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw" Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.158116 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.196875 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa121055-74c0-44aa-8401-3a8476c7db4c-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-xsj4b\" (UID: \"fa121055-74c0-44aa-8401-3a8476c7db4c\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b" Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.200653 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa121055-74c0-44aa-8401-3a8476c7db4c-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-xsj4b\" (UID: \"fa121055-74c0-44aa-8401-3a8476c7db4c\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b" Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.362921 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw" Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.388924 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-j7lcc" event={"ID":"ed23de21-61e1-463e-9a8e-5a8ba51cc727","Type":"ContainerStarted","Data":"27ab7e2bc18ae857ba2b9270ceb85efaeb2c49d460f8d8271853b9527dcd339f"} Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.390434 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-67rpb" event={"ID":"d04c81f8-7543-44b9-abc9-b0c95346cd56","Type":"ContainerStarted","Data":"b62f9bea8fa64afc8727a0bbb166fc825676fb6b1241c9d1f3365b52c65505a2"} Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.393183 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-594dc658c5-85bst"] Dec 01 17:37:06 crc kubenswrapper[4868]: W1201 17:37:06.437096 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55d8c7ab_9ae4_4313_a67c_9ef20d3e7d58.slice/crio-5615326902ae84d0a537af4eaf6c449df81f308ec2160d32135eb8430abe356c WatchSource:0}: Error finding container 5615326902ae84d0a537af4eaf6c449df81f308ec2160d32135eb8430abe356c: Status 404 returned error can't find the container with id 5615326902ae84d0a537af4eaf6c449df81f308ec2160d32135eb8430abe356c Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.470496 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b" Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.655878 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw"] Dec 01 17:37:06 crc kubenswrapper[4868]: I1201 17:37:06.732836 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b"] Dec 01 17:37:07 crc kubenswrapper[4868]: I1201 17:37:07.399377 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw" event={"ID":"f1784373-a1b0-4ca1-8c08-cf01fbc83b72","Type":"ContainerStarted","Data":"6b540da9eae976839f9a73aad36038f58806b46768c22a39b0f0fa36acc19266"} Dec 01 17:37:07 crc kubenswrapper[4868]: I1201 17:37:07.401283 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b" event={"ID":"fa121055-74c0-44aa-8401-3a8476c7db4c","Type":"ContainerStarted","Data":"b65099696202fa491467b1ce223f839246abf213a5e981ee9e6306ce1a5b0ae2"} Dec 01 17:37:07 crc kubenswrapper[4868]: I1201 17:37:07.403733 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-594dc658c5-85bst" event={"ID":"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58","Type":"ContainerStarted","Data":"d82c59b805d78a75f965d2f0547eb75021cf2f31a44ce8c0524980e9aa4bb2bb"} Dec 01 17:37:07 crc kubenswrapper[4868]: I1201 17:37:07.403772 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-594dc658c5-85bst" event={"ID":"55d8c7ab-9ae4-4313-a67c-9ef20d3e7d58","Type":"ContainerStarted","Data":"5615326902ae84d0a537af4eaf6c449df81f308ec2160d32135eb8430abe356c"} Dec 01 17:37:07 crc kubenswrapper[4868]: I1201 17:37:07.435639 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-594dc658c5-85bst" podStartSLOduration=2.43560585 podStartE2EDuration="2.43560585s" podCreationTimestamp="2025-12-01 17:37:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:37:07.426229557 +0000 UTC m=+699.797339968" watchObservedRunningTime="2025-12-01 17:37:07.43560585 +0000 UTC m=+699.806716271" Dec 01 17:37:10 crc kubenswrapper[4868]: I1201 17:37:10.434921 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-67rpb" event={"ID":"d04c81f8-7543-44b9-abc9-b0c95346cd56","Type":"ContainerStarted","Data":"adfd198314fac58e00f526b62f6626f657253dd1e2be55e9185f89e7e21898a0"} Dec 01 17:37:10 crc kubenswrapper[4868]: I1201 17:37:10.435868 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:10 crc kubenswrapper[4868]: I1201 17:37:10.439548 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw" event={"ID":"f1784373-a1b0-4ca1-8c08-cf01fbc83b72","Type":"ContainerStarted","Data":"628a584bc52e68131d34238a88d65d8f7aa5b8d8e2e2c8223099759dece5bb90"} Dec 01 17:37:10 crc kubenswrapper[4868]: I1201 17:37:10.439713 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw" Dec 01 17:37:10 crc kubenswrapper[4868]: I1201 17:37:10.443238 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-j7lcc" event={"ID":"ed23de21-61e1-463e-9a8e-5a8ba51cc727","Type":"ContainerStarted","Data":"4ed5c30b40f97ae0c0077573bf649f2e22b7999509b4d0746590fdf346278a6d"} Dec 01 17:37:10 crc kubenswrapper[4868]: I1201 17:37:10.445537 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b" event={"ID":"fa121055-74c0-44aa-8401-3a8476c7db4c","Type":"ContainerStarted","Data":"2cabd38e108428479cf6de2bd1e86f7c168b3552fbdab5375917ac258ac68eff"} Dec 01 17:37:10 crc kubenswrapper[4868]: I1201 17:37:10.471503 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-67rpb" podStartSLOduration=1.754805043 podStartE2EDuration="5.471468127s" podCreationTimestamp="2025-12-01 17:37:05 +0000 UTC" firstStartedPulling="2025-12-01 17:37:05.814266823 +0000 UTC m=+698.185377234" lastFinishedPulling="2025-12-01 17:37:09.530929887 +0000 UTC m=+701.902040318" observedRunningTime="2025-12-01 17:37:10.456622701 +0000 UTC m=+702.827733162" watchObservedRunningTime="2025-12-01 17:37:10.471468127 +0000 UTC m=+702.842578568" Dec 01 17:37:10 crc kubenswrapper[4868]: I1201 17:37:10.479485 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xsj4b" podStartSLOduration=2.72190755 podStartE2EDuration="5.479452702s" podCreationTimestamp="2025-12-01 17:37:05 +0000 UTC" firstStartedPulling="2025-12-01 17:37:06.742258081 +0000 UTC m=+699.113368482" lastFinishedPulling="2025-12-01 17:37:09.499803203 +0000 UTC m=+701.870913634" observedRunningTime="2025-12-01 17:37:10.476605722 +0000 UTC m=+702.847716143" watchObservedRunningTime="2025-12-01 17:37:10.479452702 +0000 UTC m=+702.850563123" Dec 01 17:37:10 crc kubenswrapper[4868]: I1201 17:37:10.501889 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw" podStartSLOduration=2.670500765 podStartE2EDuration="5.501866551s" podCreationTimestamp="2025-12-01 17:37:05 +0000 UTC" firstStartedPulling="2025-12-01 17:37:06.676304438 +0000 UTC m=+699.047414849" lastFinishedPulling="2025-12-01 17:37:09.507670224 +0000 UTC m=+701.878780635" observedRunningTime="2025-12-01 17:37:10.498887228 +0000 UTC m=+702.869997659" watchObservedRunningTime="2025-12-01 17:37:10.501866551 +0000 UTC m=+702.872976962" Dec 01 17:37:12 crc kubenswrapper[4868]: I1201 17:37:12.464733 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-j7lcc" event={"ID":"ed23de21-61e1-463e-9a8e-5a8ba51cc727","Type":"ContainerStarted","Data":"f2238acacf98a55580a9cf446e71e5703cb55dba58cf48053a0cfec9e2c003a8"} Dec 01 17:37:12 crc kubenswrapper[4868]: I1201 17:37:12.483819 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-j7lcc" podStartSLOduration=1.452092593 podStartE2EDuration="7.483785443s" podCreationTimestamp="2025-12-01 17:37:05 +0000 UTC" firstStartedPulling="2025-12-01 17:37:06.031481312 +0000 UTC m=+698.402591723" lastFinishedPulling="2025-12-01 17:37:12.063174162 +0000 UTC m=+704.434284573" observedRunningTime="2025-12-01 17:37:12.482805876 +0000 UTC m=+704.853916297" watchObservedRunningTime="2025-12-01 17:37:12.483785443 +0000 UTC m=+704.854895894" Dec 01 17:37:15 crc kubenswrapper[4868]: I1201 17:37:15.803574 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-67rpb" Dec 01 17:37:16 crc kubenswrapper[4868]: I1201 17:37:16.158398 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:16 crc kubenswrapper[4868]: I1201 17:37:16.158502 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:16 crc kubenswrapper[4868]: I1201 17:37:16.166629 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:16 crc kubenswrapper[4868]: I1201 17:37:16.498721 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-594dc658c5-85bst" Dec 01 17:37:16 crc kubenswrapper[4868]: I1201 17:37:16.559181 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-c7pwq"] Dec 01 17:37:26 crc kubenswrapper[4868]: I1201 17:37:26.369643 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-kjwkw" Dec 01 17:37:39 crc kubenswrapper[4868]: I1201 17:37:39.865882 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv"] Dec 01 17:37:39 crc kubenswrapper[4868]: I1201 17:37:39.868286 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" Dec 01 17:37:39 crc kubenswrapper[4868]: I1201 17:37:39.871056 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 17:37:39 crc kubenswrapper[4868]: I1201 17:37:39.877619 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv"] Dec 01 17:37:40 crc kubenswrapper[4868]: I1201 17:37:40.000790 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/22daded3-8016-41cc-bc8b-0dfc2ac3c093-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv\" (UID: \"22daded3-8016-41cc-bc8b-0dfc2ac3c093\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" Dec 01 17:37:40 crc kubenswrapper[4868]: I1201 17:37:40.001554 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96xp4\" (UniqueName: \"kubernetes.io/projected/22daded3-8016-41cc-bc8b-0dfc2ac3c093-kube-api-access-96xp4\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv\" (UID: \"22daded3-8016-41cc-bc8b-0dfc2ac3c093\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" Dec 01 17:37:40 crc kubenswrapper[4868]: I1201 17:37:40.001703 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/22daded3-8016-41cc-bc8b-0dfc2ac3c093-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv\" (UID: \"22daded3-8016-41cc-bc8b-0dfc2ac3c093\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" Dec 01 17:37:40 crc kubenswrapper[4868]: I1201 17:37:40.103046 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96xp4\" (UniqueName: \"kubernetes.io/projected/22daded3-8016-41cc-bc8b-0dfc2ac3c093-kube-api-access-96xp4\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv\" (UID: \"22daded3-8016-41cc-bc8b-0dfc2ac3c093\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" Dec 01 17:37:40 crc kubenswrapper[4868]: I1201 17:37:40.103161 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/22daded3-8016-41cc-bc8b-0dfc2ac3c093-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv\" (UID: \"22daded3-8016-41cc-bc8b-0dfc2ac3c093\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" Dec 01 17:37:40 crc kubenswrapper[4868]: I1201 17:37:40.103222 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/22daded3-8016-41cc-bc8b-0dfc2ac3c093-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv\" (UID: \"22daded3-8016-41cc-bc8b-0dfc2ac3c093\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" Dec 01 17:37:40 crc kubenswrapper[4868]: I1201 17:37:40.103758 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/22daded3-8016-41cc-bc8b-0dfc2ac3c093-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv\" (UID: \"22daded3-8016-41cc-bc8b-0dfc2ac3c093\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" Dec 01 17:37:40 crc kubenswrapper[4868]: I1201 17:37:40.103836 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/22daded3-8016-41cc-bc8b-0dfc2ac3c093-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv\" (UID: \"22daded3-8016-41cc-bc8b-0dfc2ac3c093\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" Dec 01 17:37:40 crc kubenswrapper[4868]: I1201 17:37:40.139368 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96xp4\" (UniqueName: \"kubernetes.io/projected/22daded3-8016-41cc-bc8b-0dfc2ac3c093-kube-api-access-96xp4\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv\" (UID: \"22daded3-8016-41cc-bc8b-0dfc2ac3c093\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" Dec 01 17:37:40 crc kubenswrapper[4868]: I1201 17:37:40.190964 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" Dec 01 17:37:40 crc kubenswrapper[4868]: I1201 17:37:40.660353 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv"] Dec 01 17:37:40 crc kubenswrapper[4868]: I1201 17:37:40.688593 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" event={"ID":"22daded3-8016-41cc-bc8b-0dfc2ac3c093","Type":"ContainerStarted","Data":"ee38eae37467a0c41b9cef927bea73aa277cc2e60cdd0f2323234d6b48c56209"} Dec 01 17:37:41 crc kubenswrapper[4868]: I1201 17:37:41.616798 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-c7pwq" podUID="e90813d0-8960-450f-aaa4-047341c50947" containerName="console" containerID="cri-o://76f9542b3855fd35afaa09389ce52b6759951098da66cf38e2857b1c1414008f" gracePeriod=15 Dec 01 17:37:41 crc kubenswrapper[4868]: I1201 17:37:41.698016 4868 generic.go:334] "Generic (PLEG): container finished" podID="22daded3-8016-41cc-bc8b-0dfc2ac3c093" containerID="d8a3228aa8cf28d1656a207e962a86457c3580aa7354775b0a8b558e5e0a3128" exitCode=0 Dec 01 17:37:41 crc kubenswrapper[4868]: I1201 17:37:41.698097 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" event={"ID":"22daded3-8016-41cc-bc8b-0dfc2ac3c093","Type":"ContainerDied","Data":"d8a3228aa8cf28d1656a207e962a86457c3580aa7354775b0a8b558e5e0a3128"} Dec 01 17:37:41 crc kubenswrapper[4868]: I1201 17:37:41.969130 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-c7pwq_e90813d0-8960-450f-aaa4-047341c50947/console/0.log" Dec 01 17:37:41 crc kubenswrapper[4868]: I1201 17:37:41.969504 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.137562 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e90813d0-8960-450f-aaa4-047341c50947-console-oauth-config\") pod \"e90813d0-8960-450f-aaa4-047341c50947\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.137671 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e90813d0-8960-450f-aaa4-047341c50947-console-serving-cert\") pod \"e90813d0-8960-450f-aaa4-047341c50947\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.137750 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-console-config\") pod \"e90813d0-8960-450f-aaa4-047341c50947\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.137818 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nwzv\" (UniqueName: \"kubernetes.io/projected/e90813d0-8960-450f-aaa4-047341c50947-kube-api-access-4nwzv\") pod \"e90813d0-8960-450f-aaa4-047341c50947\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.137855 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-oauth-serving-cert\") pod \"e90813d0-8960-450f-aaa4-047341c50947\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.137910 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-trusted-ca-bundle\") pod \"e90813d0-8960-450f-aaa4-047341c50947\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.138003 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-service-ca\") pod \"e90813d0-8960-450f-aaa4-047341c50947\" (UID: \"e90813d0-8960-450f-aaa4-047341c50947\") " Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.139206 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "e90813d0-8960-450f-aaa4-047341c50947" (UID: "e90813d0-8960-450f-aaa4-047341c50947"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.139230 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-console-config" (OuterVolumeSpecName: "console-config") pod "e90813d0-8960-450f-aaa4-047341c50947" (UID: "e90813d0-8960-450f-aaa4-047341c50947"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.139616 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "e90813d0-8960-450f-aaa4-047341c50947" (UID: "e90813d0-8960-450f-aaa4-047341c50947"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.140062 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-service-ca" (OuterVolumeSpecName: "service-ca") pod "e90813d0-8960-450f-aaa4-047341c50947" (UID: "e90813d0-8960-450f-aaa4-047341c50947"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.156487 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e90813d0-8960-450f-aaa4-047341c50947-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "e90813d0-8960-450f-aaa4-047341c50947" (UID: "e90813d0-8960-450f-aaa4-047341c50947"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.156635 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e90813d0-8960-450f-aaa4-047341c50947-kube-api-access-4nwzv" (OuterVolumeSpecName: "kube-api-access-4nwzv") pod "e90813d0-8960-450f-aaa4-047341c50947" (UID: "e90813d0-8960-450f-aaa4-047341c50947"). InnerVolumeSpecName "kube-api-access-4nwzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.161109 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e90813d0-8960-450f-aaa4-047341c50947-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "e90813d0-8960-450f-aaa4-047341c50947" (UID: "e90813d0-8960-450f-aaa4-047341c50947"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.239750 4868 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.239812 4868 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.239825 4868 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e90813d0-8960-450f-aaa4-047341c50947-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.239839 4868 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e90813d0-8960-450f-aaa4-047341c50947-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.239854 4868 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-console-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.239867 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nwzv\" (UniqueName: \"kubernetes.io/projected/e90813d0-8960-450f-aaa4-047341c50947-kube-api-access-4nwzv\") on node \"crc\" DevicePath \"\"" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.239882 4868 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e90813d0-8960-450f-aaa4-047341c50947-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.709668 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-c7pwq_e90813d0-8960-450f-aaa4-047341c50947/console/0.log" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.710212 4868 generic.go:334] "Generic (PLEG): container finished" podID="e90813d0-8960-450f-aaa4-047341c50947" containerID="76f9542b3855fd35afaa09389ce52b6759951098da66cf38e2857b1c1414008f" exitCode=2 Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.710331 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-c7pwq" event={"ID":"e90813d0-8960-450f-aaa4-047341c50947","Type":"ContainerDied","Data":"76f9542b3855fd35afaa09389ce52b6759951098da66cf38e2857b1c1414008f"} Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.710371 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-c7pwq" event={"ID":"e90813d0-8960-450f-aaa4-047341c50947","Type":"ContainerDied","Data":"f57bb897b63f0c317a548208464081e0b32a139de0e8639de7432b9b76808f2c"} Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.710399 4868 scope.go:117] "RemoveContainer" containerID="76f9542b3855fd35afaa09389ce52b6759951098da66cf38e2857b1c1414008f" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.710670 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-c7pwq" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.739560 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-c7pwq"] Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.749461 4868 scope.go:117] "RemoveContainer" containerID="76f9542b3855fd35afaa09389ce52b6759951098da66cf38e2857b1c1414008f" Dec 01 17:37:42 crc kubenswrapper[4868]: E1201 17:37:42.750761 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76f9542b3855fd35afaa09389ce52b6759951098da66cf38e2857b1c1414008f\": container with ID starting with 76f9542b3855fd35afaa09389ce52b6759951098da66cf38e2857b1c1414008f not found: ID does not exist" containerID="76f9542b3855fd35afaa09389ce52b6759951098da66cf38e2857b1c1414008f" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.750796 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76f9542b3855fd35afaa09389ce52b6759951098da66cf38e2857b1c1414008f"} err="failed to get container status \"76f9542b3855fd35afaa09389ce52b6759951098da66cf38e2857b1c1414008f\": rpc error: code = NotFound desc = could not find container \"76f9542b3855fd35afaa09389ce52b6759951098da66cf38e2857b1c1414008f\": container with ID starting with 76f9542b3855fd35afaa09389ce52b6759951098da66cf38e2857b1c1414008f not found: ID does not exist" Dec 01 17:37:42 crc kubenswrapper[4868]: I1201 17:37:42.752785 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-c7pwq"] Dec 01 17:37:43 crc kubenswrapper[4868]: I1201 17:37:43.718044 4868 generic.go:334] "Generic (PLEG): container finished" podID="22daded3-8016-41cc-bc8b-0dfc2ac3c093" containerID="d3728e3e51b094efc87aa3f6f6cba2df2cdf68f29071a2966af29b5a90ac8da0" exitCode=0 Dec 01 17:37:43 crc kubenswrapper[4868]: I1201 17:37:43.718101 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" event={"ID":"22daded3-8016-41cc-bc8b-0dfc2ac3c093","Type":"ContainerDied","Data":"d3728e3e51b094efc87aa3f6f6cba2df2cdf68f29071a2966af29b5a90ac8da0"} Dec 01 17:37:44 crc kubenswrapper[4868]: I1201 17:37:44.179099 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e90813d0-8960-450f-aaa4-047341c50947" path="/var/lib/kubelet/pods/e90813d0-8960-450f-aaa4-047341c50947/volumes" Dec 01 17:37:44 crc kubenswrapper[4868]: I1201 17:37:44.728973 4868 generic.go:334] "Generic (PLEG): container finished" podID="22daded3-8016-41cc-bc8b-0dfc2ac3c093" containerID="873fc62edeaed70946a10b6f645cea20f2827e8c7ff010869645edf8e0990044" exitCode=0 Dec 01 17:37:44 crc kubenswrapper[4868]: I1201 17:37:44.729037 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" event={"ID":"22daded3-8016-41cc-bc8b-0dfc2ac3c093","Type":"ContainerDied","Data":"873fc62edeaed70946a10b6f645cea20f2827e8c7ff010869645edf8e0990044"} Dec 01 17:37:46 crc kubenswrapper[4868]: I1201 17:37:46.077183 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" Dec 01 17:37:46 crc kubenswrapper[4868]: I1201 17:37:46.201521 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/22daded3-8016-41cc-bc8b-0dfc2ac3c093-util\") pod \"22daded3-8016-41cc-bc8b-0dfc2ac3c093\" (UID: \"22daded3-8016-41cc-bc8b-0dfc2ac3c093\") " Dec 01 17:37:46 crc kubenswrapper[4868]: I1201 17:37:46.201582 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/22daded3-8016-41cc-bc8b-0dfc2ac3c093-bundle\") pod \"22daded3-8016-41cc-bc8b-0dfc2ac3c093\" (UID: \"22daded3-8016-41cc-bc8b-0dfc2ac3c093\") " Dec 01 17:37:46 crc kubenswrapper[4868]: I1201 17:37:46.202280 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96xp4\" (UniqueName: \"kubernetes.io/projected/22daded3-8016-41cc-bc8b-0dfc2ac3c093-kube-api-access-96xp4\") pod \"22daded3-8016-41cc-bc8b-0dfc2ac3c093\" (UID: \"22daded3-8016-41cc-bc8b-0dfc2ac3c093\") " Dec 01 17:37:46 crc kubenswrapper[4868]: I1201 17:37:46.203416 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22daded3-8016-41cc-bc8b-0dfc2ac3c093-bundle" (OuterVolumeSpecName: "bundle") pod "22daded3-8016-41cc-bc8b-0dfc2ac3c093" (UID: "22daded3-8016-41cc-bc8b-0dfc2ac3c093"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:37:46 crc kubenswrapper[4868]: I1201 17:37:46.210829 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22daded3-8016-41cc-bc8b-0dfc2ac3c093-kube-api-access-96xp4" (OuterVolumeSpecName: "kube-api-access-96xp4") pod "22daded3-8016-41cc-bc8b-0dfc2ac3c093" (UID: "22daded3-8016-41cc-bc8b-0dfc2ac3c093"). InnerVolumeSpecName "kube-api-access-96xp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:37:46 crc kubenswrapper[4868]: I1201 17:37:46.303841 4868 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/22daded3-8016-41cc-bc8b-0dfc2ac3c093-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:37:46 crc kubenswrapper[4868]: I1201 17:37:46.303889 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96xp4\" (UniqueName: \"kubernetes.io/projected/22daded3-8016-41cc-bc8b-0dfc2ac3c093-kube-api-access-96xp4\") on node \"crc\" DevicePath \"\"" Dec 01 17:37:46 crc kubenswrapper[4868]: I1201 17:37:46.421921 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22daded3-8016-41cc-bc8b-0dfc2ac3c093-util" (OuterVolumeSpecName: "util") pod "22daded3-8016-41cc-bc8b-0dfc2ac3c093" (UID: "22daded3-8016-41cc-bc8b-0dfc2ac3c093"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:37:46 crc kubenswrapper[4868]: I1201 17:37:46.507408 4868 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/22daded3-8016-41cc-bc8b-0dfc2ac3c093-util\") on node \"crc\" DevicePath \"\"" Dec 01 17:37:46 crc kubenswrapper[4868]: I1201 17:37:46.748971 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" event={"ID":"22daded3-8016-41cc-bc8b-0dfc2ac3c093","Type":"ContainerDied","Data":"ee38eae37467a0c41b9cef927bea73aa277cc2e60cdd0f2323234d6b48c56209"} Dec 01 17:37:46 crc kubenswrapper[4868]: I1201 17:37:46.749032 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee38eae37467a0c41b9cef927bea73aa277cc2e60cdd0f2323234d6b48c56209" Dec 01 17:37:46 crc kubenswrapper[4868]: I1201 17:37:46.749124 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.146953 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-779b644dbb-f6925"] Dec 01 17:37:55 crc kubenswrapper[4868]: E1201 17:37:55.148013 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22daded3-8016-41cc-bc8b-0dfc2ac3c093" containerName="pull" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.148030 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="22daded3-8016-41cc-bc8b-0dfc2ac3c093" containerName="pull" Dec 01 17:37:55 crc kubenswrapper[4868]: E1201 17:37:55.148043 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22daded3-8016-41cc-bc8b-0dfc2ac3c093" containerName="extract" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.148049 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="22daded3-8016-41cc-bc8b-0dfc2ac3c093" containerName="extract" Dec 01 17:37:55 crc kubenswrapper[4868]: E1201 17:37:55.148056 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e90813d0-8960-450f-aaa4-047341c50947" containerName="console" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.148062 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="e90813d0-8960-450f-aaa4-047341c50947" containerName="console" Dec 01 17:37:55 crc kubenswrapper[4868]: E1201 17:37:55.148072 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22daded3-8016-41cc-bc8b-0dfc2ac3c093" containerName="util" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.148078 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="22daded3-8016-41cc-bc8b-0dfc2ac3c093" containerName="util" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.148192 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="22daded3-8016-41cc-bc8b-0dfc2ac3c093" containerName="extract" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.148204 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="e90813d0-8960-450f-aaa4-047341c50947" containerName="console" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.148679 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.157505 4868 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.157770 4868 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.161271 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.161638 4868 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-zqvqd" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.167025 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.187039 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-779b644dbb-f6925"] Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.231544 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0724b3ed-ae00-455d-9f48-94cd89bf56d9-webhook-cert\") pod \"metallb-operator-controller-manager-779b644dbb-f6925\" (UID: \"0724b3ed-ae00-455d-9f48-94cd89bf56d9\") " pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.232020 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0724b3ed-ae00-455d-9f48-94cd89bf56d9-apiservice-cert\") pod \"metallb-operator-controller-manager-779b644dbb-f6925\" (UID: \"0724b3ed-ae00-455d-9f48-94cd89bf56d9\") " pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.232144 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbw94\" (UniqueName: \"kubernetes.io/projected/0724b3ed-ae00-455d-9f48-94cd89bf56d9-kube-api-access-zbw94\") pod \"metallb-operator-controller-manager-779b644dbb-f6925\" (UID: \"0724b3ed-ae00-455d-9f48-94cd89bf56d9\") " pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.333256 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0724b3ed-ae00-455d-9f48-94cd89bf56d9-apiservice-cert\") pod \"metallb-operator-controller-manager-779b644dbb-f6925\" (UID: \"0724b3ed-ae00-455d-9f48-94cd89bf56d9\") " pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.333348 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbw94\" (UniqueName: \"kubernetes.io/projected/0724b3ed-ae00-455d-9f48-94cd89bf56d9-kube-api-access-zbw94\") pod \"metallb-operator-controller-manager-779b644dbb-f6925\" (UID: \"0724b3ed-ae00-455d-9f48-94cd89bf56d9\") " pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.333412 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0724b3ed-ae00-455d-9f48-94cd89bf56d9-webhook-cert\") pod \"metallb-operator-controller-manager-779b644dbb-f6925\" (UID: \"0724b3ed-ae00-455d-9f48-94cd89bf56d9\") " pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.341825 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0724b3ed-ae00-455d-9f48-94cd89bf56d9-apiservice-cert\") pod \"metallb-operator-controller-manager-779b644dbb-f6925\" (UID: \"0724b3ed-ae00-455d-9f48-94cd89bf56d9\") " pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.355021 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0724b3ed-ae00-455d-9f48-94cd89bf56d9-webhook-cert\") pod \"metallb-operator-controller-manager-779b644dbb-f6925\" (UID: \"0724b3ed-ae00-455d-9f48-94cd89bf56d9\") " pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.429693 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbw94\" (UniqueName: \"kubernetes.io/projected/0724b3ed-ae00-455d-9f48-94cd89bf56d9-kube-api-access-zbw94\") pod \"metallb-operator-controller-manager-779b644dbb-f6925\" (UID: \"0724b3ed-ae00-455d-9f48-94cd89bf56d9\") " pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.468727 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.623684 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6"] Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.624846 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.627242 4868 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.627616 4868 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-bw7bs" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.629224 4868 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.643059 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6"] Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.739812 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f0ad7ab5-3180-42c7-90a1-7a75fac7a61d-apiservice-cert\") pod \"metallb-operator-webhook-server-85f6846bc9-2b4v6\" (UID: \"f0ad7ab5-3180-42c7-90a1-7a75fac7a61d\") " pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.739871 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f0ad7ab5-3180-42c7-90a1-7a75fac7a61d-webhook-cert\") pod \"metallb-operator-webhook-server-85f6846bc9-2b4v6\" (UID: \"f0ad7ab5-3180-42c7-90a1-7a75fac7a61d\") " pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.740100 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48d42\" (UniqueName: \"kubernetes.io/projected/f0ad7ab5-3180-42c7-90a1-7a75fac7a61d-kube-api-access-48d42\") pod \"metallb-operator-webhook-server-85f6846bc9-2b4v6\" (UID: \"f0ad7ab5-3180-42c7-90a1-7a75fac7a61d\") " pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.842025 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f0ad7ab5-3180-42c7-90a1-7a75fac7a61d-apiservice-cert\") pod \"metallb-operator-webhook-server-85f6846bc9-2b4v6\" (UID: \"f0ad7ab5-3180-42c7-90a1-7a75fac7a61d\") " pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.842077 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f0ad7ab5-3180-42c7-90a1-7a75fac7a61d-webhook-cert\") pod \"metallb-operator-webhook-server-85f6846bc9-2b4v6\" (UID: \"f0ad7ab5-3180-42c7-90a1-7a75fac7a61d\") " pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.842131 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48d42\" (UniqueName: \"kubernetes.io/projected/f0ad7ab5-3180-42c7-90a1-7a75fac7a61d-kube-api-access-48d42\") pod \"metallb-operator-webhook-server-85f6846bc9-2b4v6\" (UID: \"f0ad7ab5-3180-42c7-90a1-7a75fac7a61d\") " pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.848812 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f0ad7ab5-3180-42c7-90a1-7a75fac7a61d-apiservice-cert\") pod \"metallb-operator-webhook-server-85f6846bc9-2b4v6\" (UID: \"f0ad7ab5-3180-42c7-90a1-7a75fac7a61d\") " pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.849786 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f0ad7ab5-3180-42c7-90a1-7a75fac7a61d-webhook-cert\") pod \"metallb-operator-webhook-server-85f6846bc9-2b4v6\" (UID: \"f0ad7ab5-3180-42c7-90a1-7a75fac7a61d\") " pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.859893 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48d42\" (UniqueName: \"kubernetes.io/projected/f0ad7ab5-3180-42c7-90a1-7a75fac7a61d-kube-api-access-48d42\") pod \"metallb-operator-webhook-server-85f6846bc9-2b4v6\" (UID: \"f0ad7ab5-3180-42c7-90a1-7a75fac7a61d\") " pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.915893 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-779b644dbb-f6925"] Dec 01 17:37:55 crc kubenswrapper[4868]: I1201 17:37:55.949435 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" Dec 01 17:37:56 crc kubenswrapper[4868]: I1201 17:37:56.273250 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6"] Dec 01 17:37:56 crc kubenswrapper[4868]: I1201 17:37:56.815504 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" event={"ID":"f0ad7ab5-3180-42c7-90a1-7a75fac7a61d","Type":"ContainerStarted","Data":"0aa5efbedf428cafe3ddf69bd2de7777daef93182b1680d6fda8a2cd60bca59f"} Dec 01 17:37:56 crc kubenswrapper[4868]: I1201 17:37:56.817005 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" event={"ID":"0724b3ed-ae00-455d-9f48-94cd89bf56d9","Type":"ContainerStarted","Data":"9e62e0a357b267099fafb871ee48bd1ee880c17a341d5025b6e79feaa7967b44"} Dec 01 17:38:03 crc kubenswrapper[4868]: I1201 17:38:03.917646 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" event={"ID":"0724b3ed-ae00-455d-9f48-94cd89bf56d9","Type":"ContainerStarted","Data":"d20bcea19a5301e1e30069f80cf7486132e7d9378e5da6d7bb42e1042cdcd97c"} Dec 01 17:38:03 crc kubenswrapper[4868]: I1201 17:38:03.918614 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" Dec 01 17:38:03 crc kubenswrapper[4868]: I1201 17:38:03.919931 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" event={"ID":"f0ad7ab5-3180-42c7-90a1-7a75fac7a61d","Type":"ContainerStarted","Data":"3cf8aacb4f9d3d34eed1110963b5bf19c7931385350f3f9a1546d2999f35d21a"} Dec 01 17:38:03 crc kubenswrapper[4868]: I1201 17:38:03.920376 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" Dec 01 17:38:03 crc kubenswrapper[4868]: I1201 17:38:03.941265 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" podStartSLOduration=1.502352176 podStartE2EDuration="8.941237734s" podCreationTimestamp="2025-12-01 17:37:55 +0000 UTC" firstStartedPulling="2025-12-01 17:37:55.936495698 +0000 UTC m=+748.307606109" lastFinishedPulling="2025-12-01 17:38:03.375381256 +0000 UTC m=+755.746491667" observedRunningTime="2025-12-01 17:38:03.936777669 +0000 UTC m=+756.307888090" watchObservedRunningTime="2025-12-01 17:38:03.941237734 +0000 UTC m=+756.312348145" Dec 01 17:38:03 crc kubenswrapper[4868]: I1201 17:38:03.964755 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" podStartSLOduration=1.856060689 podStartE2EDuration="8.964726137s" podCreationTimestamp="2025-12-01 17:37:55 +0000 UTC" firstStartedPulling="2025-12-01 17:37:56.281048777 +0000 UTC m=+748.652159188" lastFinishedPulling="2025-12-01 17:38:03.389714225 +0000 UTC m=+755.760824636" observedRunningTime="2025-12-01 17:38:03.960588032 +0000 UTC m=+756.331698443" watchObservedRunningTime="2025-12-01 17:38:03.964726137 +0000 UTC m=+756.335836558" Dec 01 17:38:12 crc kubenswrapper[4868]: I1201 17:38:12.791077 4868 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 17:38:15 crc kubenswrapper[4868]: I1201 17:38:15.956283 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-85f6846bc9-2b4v6" Dec 01 17:38:25 crc kubenswrapper[4868]: I1201 17:38:25.904376 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:38:25 crc kubenswrapper[4868]: I1201 17:38:25.905213 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:38:35 crc kubenswrapper[4868]: I1201 17:38:35.471885 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-779b644dbb-f6925" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.253273 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-57x2t"] Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.256192 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.257669 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t"] Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.258816 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.259110 4868 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-pk5lg" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.259170 4868 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.261005 4868 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.261098 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.299702 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t"] Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.399737 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hgpl\" (UniqueName: \"kubernetes.io/projected/e22c99ca-b516-4e36-bc9b-ba5355170fd8-kube-api-access-4hgpl\") pod \"frr-k8s-webhook-server-7fcb986d4-54t6t\" (UID: \"e22c99ca-b516-4e36-bc9b-ba5355170fd8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.399796 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-frr-conf\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.399844 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-metrics\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.400037 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-427x9\" (UniqueName: \"kubernetes.io/projected/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-kube-api-access-427x9\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.400230 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-frr-sockets\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.400257 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-metrics-certs\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.400276 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e22c99ca-b516-4e36-bc9b-ba5355170fd8-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-54t6t\" (UID: \"e22c99ca-b516-4e36-bc9b-ba5355170fd8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.400297 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-reloader\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.400392 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-frr-startup\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.425800 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-xfljl"] Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.427632 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-xfljl" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.431549 4868 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.432005 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.432098 4868 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-6f5hz" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.432404 4868 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.443115 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-nvnm5"] Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.444140 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-nvnm5" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.445741 4868 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.468256 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-nvnm5"] Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.501673 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-metrics\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.501771 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-427x9\" (UniqueName: \"kubernetes.io/projected/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-kube-api-access-427x9\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.501838 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-frr-sockets\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.501868 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-metrics-certs\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.501902 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e22c99ca-b516-4e36-bc9b-ba5355170fd8-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-54t6t\" (UID: \"e22c99ca-b516-4e36-bc9b-ba5355170fd8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.501920 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-reloader\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.501962 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-frr-startup\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.501987 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hgpl\" (UniqueName: \"kubernetes.io/projected/e22c99ca-b516-4e36-bc9b-ba5355170fd8-kube-api-access-4hgpl\") pod \"frr-k8s-webhook-server-7fcb986d4-54t6t\" (UID: \"e22c99ca-b516-4e36-bc9b-ba5355170fd8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.502006 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-frr-conf\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.503636 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-metrics\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.503790 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-frr-sockets\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.503880 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-frr-conf\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.504214 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-reloader\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.504702 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-frr-startup\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.516043 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e22c99ca-b516-4e36-bc9b-ba5355170fd8-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-54t6t\" (UID: \"e22c99ca-b516-4e36-bc9b-ba5355170fd8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.523450 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-metrics-certs\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.529471 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hgpl\" (UniqueName: \"kubernetes.io/projected/e22c99ca-b516-4e36-bc9b-ba5355170fd8-kube-api-access-4hgpl\") pod \"frr-k8s-webhook-server-7fcb986d4-54t6t\" (UID: \"e22c99ca-b516-4e36-bc9b-ba5355170fd8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.532902 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-427x9\" (UniqueName: \"kubernetes.io/projected/86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41-kube-api-access-427x9\") pod \"frr-k8s-57x2t\" (UID: \"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41\") " pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.590084 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.601538 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.603142 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5536b1ca-ecee-4f69-9884-539b000d00f0-metrics-certs\") pod \"controller-f8648f98b-nvnm5\" (UID: \"5536b1ca-ecee-4f69-9884-539b000d00f0\") " pod="metallb-system/controller-f8648f98b-nvnm5" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.603189 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/22663c8b-3ec6-453c-bb25-ca8d92576d07-metallb-excludel2\") pod \"speaker-xfljl\" (UID: \"22663c8b-3ec6-453c-bb25-ca8d92576d07\") " pod="metallb-system/speaker-xfljl" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.603225 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22663c8b-3ec6-453c-bb25-ca8d92576d07-metrics-certs\") pod \"speaker-xfljl\" (UID: \"22663c8b-3ec6-453c-bb25-ca8d92576d07\") " pod="metallb-system/speaker-xfljl" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.603243 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tmk9\" (UniqueName: \"kubernetes.io/projected/22663c8b-3ec6-453c-bb25-ca8d92576d07-kube-api-access-4tmk9\") pod \"speaker-xfljl\" (UID: \"22663c8b-3ec6-453c-bb25-ca8d92576d07\") " pod="metallb-system/speaker-xfljl" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.603329 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/22663c8b-3ec6-453c-bb25-ca8d92576d07-memberlist\") pod \"speaker-xfljl\" (UID: \"22663c8b-3ec6-453c-bb25-ca8d92576d07\") " pod="metallb-system/speaker-xfljl" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.603390 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5536b1ca-ecee-4f69-9884-539b000d00f0-cert\") pod \"controller-f8648f98b-nvnm5\" (UID: \"5536b1ca-ecee-4f69-9884-539b000d00f0\") " pod="metallb-system/controller-f8648f98b-nvnm5" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.603443 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8kdt\" (UniqueName: \"kubernetes.io/projected/5536b1ca-ecee-4f69-9884-539b000d00f0-kube-api-access-r8kdt\") pod \"controller-f8648f98b-nvnm5\" (UID: \"5536b1ca-ecee-4f69-9884-539b000d00f0\") " pod="metallb-system/controller-f8648f98b-nvnm5" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.704826 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/22663c8b-3ec6-453c-bb25-ca8d92576d07-memberlist\") pod \"speaker-xfljl\" (UID: \"22663c8b-3ec6-453c-bb25-ca8d92576d07\") " pod="metallb-system/speaker-xfljl" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.704883 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5536b1ca-ecee-4f69-9884-539b000d00f0-cert\") pod \"controller-f8648f98b-nvnm5\" (UID: \"5536b1ca-ecee-4f69-9884-539b000d00f0\") " pod="metallb-system/controller-f8648f98b-nvnm5" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.704913 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8kdt\" (UniqueName: \"kubernetes.io/projected/5536b1ca-ecee-4f69-9884-539b000d00f0-kube-api-access-r8kdt\") pod \"controller-f8648f98b-nvnm5\" (UID: \"5536b1ca-ecee-4f69-9884-539b000d00f0\") " pod="metallb-system/controller-f8648f98b-nvnm5" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.704973 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5536b1ca-ecee-4f69-9884-539b000d00f0-metrics-certs\") pod \"controller-f8648f98b-nvnm5\" (UID: \"5536b1ca-ecee-4f69-9884-539b000d00f0\") " pod="metallb-system/controller-f8648f98b-nvnm5" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.704996 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/22663c8b-3ec6-453c-bb25-ca8d92576d07-metallb-excludel2\") pod \"speaker-xfljl\" (UID: \"22663c8b-3ec6-453c-bb25-ca8d92576d07\") " pod="metallb-system/speaker-xfljl" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.705013 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22663c8b-3ec6-453c-bb25-ca8d92576d07-metrics-certs\") pod \"speaker-xfljl\" (UID: \"22663c8b-3ec6-453c-bb25-ca8d92576d07\") " pod="metallb-system/speaker-xfljl" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.705028 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tmk9\" (UniqueName: \"kubernetes.io/projected/22663c8b-3ec6-453c-bb25-ca8d92576d07-kube-api-access-4tmk9\") pod \"speaker-xfljl\" (UID: \"22663c8b-3ec6-453c-bb25-ca8d92576d07\") " pod="metallb-system/speaker-xfljl" Dec 01 17:38:36 crc kubenswrapper[4868]: E1201 17:38:36.705158 4868 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 17:38:36 crc kubenswrapper[4868]: E1201 17:38:36.705271 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22663c8b-3ec6-453c-bb25-ca8d92576d07-memberlist podName:22663c8b-3ec6-453c-bb25-ca8d92576d07 nodeName:}" failed. No retries permitted until 2025-12-01 17:38:37.205244348 +0000 UTC m=+789.576354949 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/22663c8b-3ec6-453c-bb25-ca8d92576d07-memberlist") pod "speaker-xfljl" (UID: "22663c8b-3ec6-453c-bb25-ca8d92576d07") : secret "metallb-memberlist" not found Dec 01 17:38:36 crc kubenswrapper[4868]: E1201 17:38:36.705837 4868 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 01 17:38:36 crc kubenswrapper[4868]: E1201 17:38:36.705889 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22663c8b-3ec6-453c-bb25-ca8d92576d07-metrics-certs podName:22663c8b-3ec6-453c-bb25-ca8d92576d07 nodeName:}" failed. No retries permitted until 2025-12-01 17:38:37.205876495 +0000 UTC m=+789.576986906 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/22663c8b-3ec6-453c-bb25-ca8d92576d07-metrics-certs") pod "speaker-xfljl" (UID: "22663c8b-3ec6-453c-bb25-ca8d92576d07") : secret "speaker-certs-secret" not found Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.706477 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/22663c8b-3ec6-453c-bb25-ca8d92576d07-metallb-excludel2\") pod \"speaker-xfljl\" (UID: \"22663c8b-3ec6-453c-bb25-ca8d92576d07\") " pod="metallb-system/speaker-xfljl" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.713153 4868 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.731513 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5536b1ca-ecee-4f69-9884-539b000d00f0-cert\") pod \"controller-f8648f98b-nvnm5\" (UID: \"5536b1ca-ecee-4f69-9884-539b000d00f0\") " pod="metallb-system/controller-f8648f98b-nvnm5" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.735963 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5536b1ca-ecee-4f69-9884-539b000d00f0-metrics-certs\") pod \"controller-f8648f98b-nvnm5\" (UID: \"5536b1ca-ecee-4f69-9884-539b000d00f0\") " pod="metallb-system/controller-f8648f98b-nvnm5" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.736683 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tmk9\" (UniqueName: \"kubernetes.io/projected/22663c8b-3ec6-453c-bb25-ca8d92576d07-kube-api-access-4tmk9\") pod \"speaker-xfljl\" (UID: \"22663c8b-3ec6-453c-bb25-ca8d92576d07\") " pod="metallb-system/speaker-xfljl" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.743708 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8kdt\" (UniqueName: \"kubernetes.io/projected/5536b1ca-ecee-4f69-9884-539b000d00f0-kube-api-access-r8kdt\") pod \"controller-f8648f98b-nvnm5\" (UID: \"5536b1ca-ecee-4f69-9884-539b000d00f0\") " pod="metallb-system/controller-f8648f98b-nvnm5" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.762858 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-nvnm5" Dec 01 17:38:36 crc kubenswrapper[4868]: I1201 17:38:36.957032 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-nvnm5"] Dec 01 17:38:36 crc kubenswrapper[4868]: W1201 17:38:36.965161 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5536b1ca_ecee_4f69_9884_539b000d00f0.slice/crio-6d6e6b1bd58c9063528e7acd858ad36d4fe3100e802cd70604ed0d98348df599 WatchSource:0}: Error finding container 6d6e6b1bd58c9063528e7acd858ad36d4fe3100e802cd70604ed0d98348df599: Status 404 returned error can't find the container with id 6d6e6b1bd58c9063528e7acd858ad36d4fe3100e802cd70604ed0d98348df599 Dec 01 17:38:37 crc kubenswrapper[4868]: I1201 17:38:37.115591 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t"] Dec 01 17:38:37 crc kubenswrapper[4868]: W1201 17:38:37.122663 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode22c99ca_b516_4e36_bc9b_ba5355170fd8.slice/crio-2fa22c2753492444ebbde9e59b7a60bac8bf060c3a9d81c05f9d88699c509cf3 WatchSource:0}: Error finding container 2fa22c2753492444ebbde9e59b7a60bac8bf060c3a9d81c05f9d88699c509cf3: Status 404 returned error can't find the container with id 2fa22c2753492444ebbde9e59b7a60bac8bf060c3a9d81c05f9d88699c509cf3 Dec 01 17:38:37 crc kubenswrapper[4868]: I1201 17:38:37.152655 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t" event={"ID":"e22c99ca-b516-4e36-bc9b-ba5355170fd8","Type":"ContainerStarted","Data":"2fa22c2753492444ebbde9e59b7a60bac8bf060c3a9d81c05f9d88699c509cf3"} Dec 01 17:38:37 crc kubenswrapper[4868]: I1201 17:38:37.154164 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-57x2t" event={"ID":"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41","Type":"ContainerStarted","Data":"bd38e31aea6aa484b3f7a098d10f68eccb260b6473c78b90e96583cae05c71d0"} Dec 01 17:38:37 crc kubenswrapper[4868]: I1201 17:38:37.160559 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-nvnm5" event={"ID":"5536b1ca-ecee-4f69-9884-539b000d00f0","Type":"ContainerStarted","Data":"445aca1fafc69cc387c2aba6165734db111d152a7651eb0ea1ecbfec6d9db195"} Dec 01 17:38:37 crc kubenswrapper[4868]: I1201 17:38:37.160614 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-nvnm5" event={"ID":"5536b1ca-ecee-4f69-9884-539b000d00f0","Type":"ContainerStarted","Data":"6d6e6b1bd58c9063528e7acd858ad36d4fe3100e802cd70604ed0d98348df599"} Dec 01 17:38:37 crc kubenswrapper[4868]: I1201 17:38:37.212671 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/22663c8b-3ec6-453c-bb25-ca8d92576d07-memberlist\") pod \"speaker-xfljl\" (UID: \"22663c8b-3ec6-453c-bb25-ca8d92576d07\") " pod="metallb-system/speaker-xfljl" Dec 01 17:38:37 crc kubenswrapper[4868]: I1201 17:38:37.212781 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22663c8b-3ec6-453c-bb25-ca8d92576d07-metrics-certs\") pod \"speaker-xfljl\" (UID: \"22663c8b-3ec6-453c-bb25-ca8d92576d07\") " pod="metallb-system/speaker-xfljl" Dec 01 17:38:37 crc kubenswrapper[4868]: E1201 17:38:37.212841 4868 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 17:38:37 crc kubenswrapper[4868]: E1201 17:38:37.212933 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22663c8b-3ec6-453c-bb25-ca8d92576d07-memberlist podName:22663c8b-3ec6-453c-bb25-ca8d92576d07 nodeName:}" failed. No retries permitted until 2025-12-01 17:38:38.212907716 +0000 UTC m=+790.584018147 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/22663c8b-3ec6-453c-bb25-ca8d92576d07-memberlist") pod "speaker-xfljl" (UID: "22663c8b-3ec6-453c-bb25-ca8d92576d07") : secret "metallb-memberlist" not found Dec 01 17:38:37 crc kubenswrapper[4868]: I1201 17:38:37.219071 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22663c8b-3ec6-453c-bb25-ca8d92576d07-metrics-certs\") pod \"speaker-xfljl\" (UID: \"22663c8b-3ec6-453c-bb25-ca8d92576d07\") " pod="metallb-system/speaker-xfljl" Dec 01 17:38:38 crc kubenswrapper[4868]: I1201 17:38:38.185486 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-nvnm5" Dec 01 17:38:38 crc kubenswrapper[4868]: I1201 17:38:38.185965 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-nvnm5" event={"ID":"5536b1ca-ecee-4f69-9884-539b000d00f0","Type":"ContainerStarted","Data":"2b1893d0d3d5db761b523d485cc69688d5eb9292d314dc5a802cb0cf44a2e676"} Dec 01 17:38:38 crc kubenswrapper[4868]: I1201 17:38:38.228066 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/22663c8b-3ec6-453c-bb25-ca8d92576d07-memberlist\") pod \"speaker-xfljl\" (UID: \"22663c8b-3ec6-453c-bb25-ca8d92576d07\") " pod="metallb-system/speaker-xfljl" Dec 01 17:38:38 crc kubenswrapper[4868]: I1201 17:38:38.235217 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/22663c8b-3ec6-453c-bb25-ca8d92576d07-memberlist\") pod \"speaker-xfljl\" (UID: \"22663c8b-3ec6-453c-bb25-ca8d92576d07\") " pod="metallb-system/speaker-xfljl" Dec 01 17:38:38 crc kubenswrapper[4868]: I1201 17:38:38.241168 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-xfljl" Dec 01 17:38:38 crc kubenswrapper[4868]: I1201 17:38:38.248592 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-nvnm5" podStartSLOduration=2.248549737 podStartE2EDuration="2.248549737s" podCreationTimestamp="2025-12-01 17:38:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:38:38.242866339 +0000 UTC m=+790.613976750" watchObservedRunningTime="2025-12-01 17:38:38.248549737 +0000 UTC m=+790.619660168" Dec 01 17:38:39 crc kubenswrapper[4868]: I1201 17:38:39.187491 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-xfljl" event={"ID":"22663c8b-3ec6-453c-bb25-ca8d92576d07","Type":"ContainerStarted","Data":"36b4a014b92ce34e63dd1a01ce040ce29fce6cee660385e7e21d33009e6b093d"} Dec 01 17:38:39 crc kubenswrapper[4868]: I1201 17:38:39.188560 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-xfljl" event={"ID":"22663c8b-3ec6-453c-bb25-ca8d92576d07","Type":"ContainerStarted","Data":"937396c8effb6c399c06f3a2bdd14cec26d134e35fb68b70edc724cd993658d2"} Dec 01 17:38:39 crc kubenswrapper[4868]: I1201 17:38:39.188579 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-xfljl" event={"ID":"22663c8b-3ec6-453c-bb25-ca8d92576d07","Type":"ContainerStarted","Data":"e6f66f70b7f5ba79045157faab6d6b19fe33cce3cffda5bc30e83a56f9396e39"} Dec 01 17:38:39 crc kubenswrapper[4868]: I1201 17:38:39.188911 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-xfljl" Dec 01 17:38:39 crc kubenswrapper[4868]: I1201 17:38:39.216108 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-xfljl" podStartSLOduration=3.216080223 podStartE2EDuration="3.216080223s" podCreationTimestamp="2025-12-01 17:38:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:38:39.208189773 +0000 UTC m=+791.579300184" watchObservedRunningTime="2025-12-01 17:38:39.216080223 +0000 UTC m=+791.587190634" Dec 01 17:38:45 crc kubenswrapper[4868]: I1201 17:38:45.244305 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t" event={"ID":"e22c99ca-b516-4e36-bc9b-ba5355170fd8","Type":"ContainerStarted","Data":"02def5a143161300a212d30b4dec495798760adc71ce3d2721c0d58a8cfd9a08"} Dec 01 17:38:45 crc kubenswrapper[4868]: I1201 17:38:45.245885 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t" Dec 01 17:38:45 crc kubenswrapper[4868]: I1201 17:38:45.247171 4868 generic.go:334] "Generic (PLEG): container finished" podID="86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41" containerID="4fb0c4aa252a73d556366d5645a48c54f6c9f62f6b5e0966d7d319b3516f8a45" exitCode=0 Dec 01 17:38:45 crc kubenswrapper[4868]: I1201 17:38:45.247280 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-57x2t" event={"ID":"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41","Type":"ContainerDied","Data":"4fb0c4aa252a73d556366d5645a48c54f6c9f62f6b5e0966d7d319b3516f8a45"} Dec 01 17:38:45 crc kubenswrapper[4868]: I1201 17:38:45.274110 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t" podStartSLOduration=1.652254742 podStartE2EDuration="9.274078002s" podCreationTimestamp="2025-12-01 17:38:36 +0000 UTC" firstStartedPulling="2025-12-01 17:38:37.127323274 +0000 UTC m=+789.498433685" lastFinishedPulling="2025-12-01 17:38:44.749146534 +0000 UTC m=+797.120256945" observedRunningTime="2025-12-01 17:38:45.269554057 +0000 UTC m=+797.640664508" watchObservedRunningTime="2025-12-01 17:38:45.274078002 +0000 UTC m=+797.645188453" Dec 01 17:38:46 crc kubenswrapper[4868]: I1201 17:38:46.256372 4868 generic.go:334] "Generic (PLEG): container finished" podID="86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41" containerID="570b0ae878e61c615f94f0b4118ba8187aec6253ccddbcac7829ab7d8d890249" exitCode=0 Dec 01 17:38:46 crc kubenswrapper[4868]: I1201 17:38:46.257847 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-57x2t" event={"ID":"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41","Type":"ContainerDied","Data":"570b0ae878e61c615f94f0b4118ba8187aec6253ccddbcac7829ab7d8d890249"} Dec 01 17:38:47 crc kubenswrapper[4868]: I1201 17:38:47.266729 4868 generic.go:334] "Generic (PLEG): container finished" podID="86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41" containerID="1368bce4da4208b68b4674585cd84f569d1e352d51a92a8f33e751cf79854d06" exitCode=0 Dec 01 17:38:47 crc kubenswrapper[4868]: I1201 17:38:47.266827 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-57x2t" event={"ID":"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41","Type":"ContainerDied","Data":"1368bce4da4208b68b4674585cd84f569d1e352d51a92a8f33e751cf79854d06"} Dec 01 17:38:48 crc kubenswrapper[4868]: I1201 17:38:48.249512 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-xfljl" Dec 01 17:38:48 crc kubenswrapper[4868]: I1201 17:38:48.299848 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-57x2t" event={"ID":"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41","Type":"ContainerStarted","Data":"c5dcfada8dae548f06c4f5a4e5870ff04caf0272d270b98e162d2f0e4b8c818a"} Dec 01 17:38:48 crc kubenswrapper[4868]: I1201 17:38:48.299914 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-57x2t" event={"ID":"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41","Type":"ContainerStarted","Data":"14fb76c147bb5fb3c181f322a5bfa8822e7e27c406889cd1ea1fc00d08230449"} Dec 01 17:38:48 crc kubenswrapper[4868]: I1201 17:38:48.299928 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-57x2t" event={"ID":"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41","Type":"ContainerStarted","Data":"89c1c912d18ad2bf10f578cb0ee176dae2ab877cd695447c4885eaf3af0451ed"} Dec 01 17:38:48 crc kubenswrapper[4868]: I1201 17:38:48.299958 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-57x2t" event={"ID":"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41","Type":"ContainerStarted","Data":"a7074618ebe7ecf646a6b5217ef61137ec7176e8247ef1919d0bf5b67b4cc4ca"} Dec 01 17:38:48 crc kubenswrapper[4868]: I1201 17:38:48.299975 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-57x2t" event={"ID":"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41","Type":"ContainerStarted","Data":"e2515529bbd2c4da826532b729658937275c3597b78ee3597973314dd30e0a8e"} Dec 01 17:38:49 crc kubenswrapper[4868]: I1201 17:38:49.312998 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-57x2t" event={"ID":"86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41","Type":"ContainerStarted","Data":"2af656e46f22b849b81567af918c48506bcb37842219c8c2a795f1d2286b5869"} Dec 01 17:38:49 crc kubenswrapper[4868]: I1201 17:38:49.314721 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:49 crc kubenswrapper[4868]: I1201 17:38:49.339606 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-57x2t" podStartSLOduration=5.534418292 podStartE2EDuration="13.339566533s" podCreationTimestamp="2025-12-01 17:38:36 +0000 UTC" firstStartedPulling="2025-12-01 17:38:36.957154539 +0000 UTC m=+789.328264950" lastFinishedPulling="2025-12-01 17:38:44.76230274 +0000 UTC m=+797.133413191" observedRunningTime="2025-12-01 17:38:49.33843068 +0000 UTC m=+801.709541111" watchObservedRunningTime="2025-12-01 17:38:49.339566533 +0000 UTC m=+801.710676994" Dec 01 17:38:51 crc kubenswrapper[4868]: I1201 17:38:51.219657 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-5clwr"] Dec 01 17:38:51 crc kubenswrapper[4868]: I1201 17:38:51.220929 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5clwr" Dec 01 17:38:51 crc kubenswrapper[4868]: I1201 17:38:51.223257 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 01 17:38:51 crc kubenswrapper[4868]: I1201 17:38:51.223319 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 01 17:38:51 crc kubenswrapper[4868]: I1201 17:38:51.224153 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-c9wnn" Dec 01 17:38:51 crc kubenswrapper[4868]: I1201 17:38:51.237479 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5clwr"] Dec 01 17:38:51 crc kubenswrapper[4868]: I1201 17:38:51.322917 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crksb\" (UniqueName: \"kubernetes.io/projected/25f64f3d-b811-4548-b12e-5d1bf8b1dc12-kube-api-access-crksb\") pod \"openstack-operator-index-5clwr\" (UID: \"25f64f3d-b811-4548-b12e-5d1bf8b1dc12\") " pod="openstack-operators/openstack-operator-index-5clwr" Dec 01 17:38:51 crc kubenswrapper[4868]: I1201 17:38:51.424527 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crksb\" (UniqueName: \"kubernetes.io/projected/25f64f3d-b811-4548-b12e-5d1bf8b1dc12-kube-api-access-crksb\") pod \"openstack-operator-index-5clwr\" (UID: \"25f64f3d-b811-4548-b12e-5d1bf8b1dc12\") " pod="openstack-operators/openstack-operator-index-5clwr" Dec 01 17:38:51 crc kubenswrapper[4868]: I1201 17:38:51.450529 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crksb\" (UniqueName: \"kubernetes.io/projected/25f64f3d-b811-4548-b12e-5d1bf8b1dc12-kube-api-access-crksb\") pod \"openstack-operator-index-5clwr\" (UID: \"25f64f3d-b811-4548-b12e-5d1bf8b1dc12\") " pod="openstack-operators/openstack-operator-index-5clwr" Dec 01 17:38:51 crc kubenswrapper[4868]: I1201 17:38:51.541786 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5clwr" Dec 01 17:38:51 crc kubenswrapper[4868]: I1201 17:38:51.591282 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:51 crc kubenswrapper[4868]: I1201 17:38:51.645049 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-57x2t" Dec 01 17:38:51 crc kubenswrapper[4868]: I1201 17:38:51.999669 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5clwr"] Dec 01 17:38:52 crc kubenswrapper[4868]: W1201 17:38:52.010119 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25f64f3d_b811_4548_b12e_5d1bf8b1dc12.slice/crio-f3d20f04bee80fe66f5f34a6883b2a2a84525bdbaa69326c0ba6ca0a74af55f9 WatchSource:0}: Error finding container f3d20f04bee80fe66f5f34a6883b2a2a84525bdbaa69326c0ba6ca0a74af55f9: Status 404 returned error can't find the container with id f3d20f04bee80fe66f5f34a6883b2a2a84525bdbaa69326c0ba6ca0a74af55f9 Dec 01 17:38:52 crc kubenswrapper[4868]: I1201 17:38:52.337031 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5clwr" event={"ID":"25f64f3d-b811-4548-b12e-5d1bf8b1dc12","Type":"ContainerStarted","Data":"f3d20f04bee80fe66f5f34a6883b2a2a84525bdbaa69326c0ba6ca0a74af55f9"} Dec 01 17:38:54 crc kubenswrapper[4868]: I1201 17:38:54.395547 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-5clwr"] Dec 01 17:38:55 crc kubenswrapper[4868]: I1201 17:38:55.002304 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-8rf2p"] Dec 01 17:38:55 crc kubenswrapper[4868]: I1201 17:38:55.003584 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8rf2p" Dec 01 17:38:55 crc kubenswrapper[4868]: I1201 17:38:55.013140 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-8rf2p"] Dec 01 17:38:55 crc kubenswrapper[4868]: I1201 17:38:55.186174 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5bxf\" (UniqueName: \"kubernetes.io/projected/d73d228d-a5b1-4569-8c1f-a062812250b9-kube-api-access-f5bxf\") pod \"openstack-operator-index-8rf2p\" (UID: \"d73d228d-a5b1-4569-8c1f-a062812250b9\") " pod="openstack-operators/openstack-operator-index-8rf2p" Dec 01 17:38:55 crc kubenswrapper[4868]: I1201 17:38:55.287969 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5bxf\" (UniqueName: \"kubernetes.io/projected/d73d228d-a5b1-4569-8c1f-a062812250b9-kube-api-access-f5bxf\") pod \"openstack-operator-index-8rf2p\" (UID: \"d73d228d-a5b1-4569-8c1f-a062812250b9\") " pod="openstack-operators/openstack-operator-index-8rf2p" Dec 01 17:38:55 crc kubenswrapper[4868]: I1201 17:38:55.312717 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5bxf\" (UniqueName: \"kubernetes.io/projected/d73d228d-a5b1-4569-8c1f-a062812250b9-kube-api-access-f5bxf\") pod \"openstack-operator-index-8rf2p\" (UID: \"d73d228d-a5b1-4569-8c1f-a062812250b9\") " pod="openstack-operators/openstack-operator-index-8rf2p" Dec 01 17:38:55 crc kubenswrapper[4868]: I1201 17:38:55.330128 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8rf2p" Dec 01 17:38:55 crc kubenswrapper[4868]: I1201 17:38:55.904740 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:38:55 crc kubenswrapper[4868]: I1201 17:38:55.905715 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:38:56 crc kubenswrapper[4868]: I1201 17:38:56.052793 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-8rf2p"] Dec 01 17:38:56 crc kubenswrapper[4868]: W1201 17:38:56.060861 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd73d228d_a5b1_4569_8c1f_a062812250b9.slice/crio-8f7f2c3f26b6cba36e097fda46da53dc1597537451be7035e9f176cddd7f738b WatchSource:0}: Error finding container 8f7f2c3f26b6cba36e097fda46da53dc1597537451be7035e9f176cddd7f738b: Status 404 returned error can't find the container with id 8f7f2c3f26b6cba36e097fda46da53dc1597537451be7035e9f176cddd7f738b Dec 01 17:38:56 crc kubenswrapper[4868]: I1201 17:38:56.365821 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5clwr" event={"ID":"25f64f3d-b811-4548-b12e-5d1bf8b1dc12","Type":"ContainerStarted","Data":"75b57ec40d32c1a005276f349c3f07e091f3b545b449617e3a9d1862a8c5619d"} Dec 01 17:38:56 crc kubenswrapper[4868]: I1201 17:38:56.365908 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-5clwr" podUID="25f64f3d-b811-4548-b12e-5d1bf8b1dc12" containerName="registry-server" containerID="cri-o://75b57ec40d32c1a005276f349c3f07e091f3b545b449617e3a9d1862a8c5619d" gracePeriod=2 Dec 01 17:38:56 crc kubenswrapper[4868]: I1201 17:38:56.370344 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8rf2p" event={"ID":"d73d228d-a5b1-4569-8c1f-a062812250b9","Type":"ContainerStarted","Data":"8ce7f18cf994b8b32899a2737a1001154b5caba07d1a7c6141d4ae4caf65acbd"} Dec 01 17:38:56 crc kubenswrapper[4868]: I1201 17:38:56.370393 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8rf2p" event={"ID":"d73d228d-a5b1-4569-8c1f-a062812250b9","Type":"ContainerStarted","Data":"8f7f2c3f26b6cba36e097fda46da53dc1597537451be7035e9f176cddd7f738b"} Dec 01 17:38:56 crc kubenswrapper[4868]: I1201 17:38:56.401540 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-8rf2p" podStartSLOduration=2.340528805 podStartE2EDuration="2.401522262s" podCreationTimestamp="2025-12-01 17:38:54 +0000 UTC" firstStartedPulling="2025-12-01 17:38:56.065631814 +0000 UTC m=+808.436742225" lastFinishedPulling="2025-12-01 17:38:56.126625271 +0000 UTC m=+808.497735682" observedRunningTime="2025-12-01 17:38:56.398852128 +0000 UTC m=+808.769962539" watchObservedRunningTime="2025-12-01 17:38:56.401522262 +0000 UTC m=+808.772632673" Dec 01 17:38:56 crc kubenswrapper[4868]: I1201 17:38:56.403360 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-5clwr" podStartSLOduration=1.746131134 podStartE2EDuration="5.403353332s" podCreationTimestamp="2025-12-01 17:38:51 +0000 UTC" firstStartedPulling="2025-12-01 17:38:52.012673753 +0000 UTC m=+804.383784164" lastFinishedPulling="2025-12-01 17:38:55.669895951 +0000 UTC m=+808.041006362" observedRunningTime="2025-12-01 17:38:56.383003586 +0000 UTC m=+808.754113997" watchObservedRunningTime="2025-12-01 17:38:56.403353332 +0000 UTC m=+808.774463743" Dec 01 17:38:56 crc kubenswrapper[4868]: I1201 17:38:56.607328 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-54t6t" Dec 01 17:38:56 crc kubenswrapper[4868]: I1201 17:38:56.723059 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5clwr" Dec 01 17:38:56 crc kubenswrapper[4868]: I1201 17:38:56.769293 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-nvnm5" Dec 01 17:38:56 crc kubenswrapper[4868]: I1201 17:38:56.914578 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crksb\" (UniqueName: \"kubernetes.io/projected/25f64f3d-b811-4548-b12e-5d1bf8b1dc12-kube-api-access-crksb\") pod \"25f64f3d-b811-4548-b12e-5d1bf8b1dc12\" (UID: \"25f64f3d-b811-4548-b12e-5d1bf8b1dc12\") " Dec 01 17:38:56 crc kubenswrapper[4868]: I1201 17:38:56.924039 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25f64f3d-b811-4548-b12e-5d1bf8b1dc12-kube-api-access-crksb" (OuterVolumeSpecName: "kube-api-access-crksb") pod "25f64f3d-b811-4548-b12e-5d1bf8b1dc12" (UID: "25f64f3d-b811-4548-b12e-5d1bf8b1dc12"). InnerVolumeSpecName "kube-api-access-crksb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:38:57 crc kubenswrapper[4868]: I1201 17:38:57.015829 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crksb\" (UniqueName: \"kubernetes.io/projected/25f64f3d-b811-4548-b12e-5d1bf8b1dc12-kube-api-access-crksb\") on node \"crc\" DevicePath \"\"" Dec 01 17:38:57 crc kubenswrapper[4868]: I1201 17:38:57.415457 4868 generic.go:334] "Generic (PLEG): container finished" podID="25f64f3d-b811-4548-b12e-5d1bf8b1dc12" containerID="75b57ec40d32c1a005276f349c3f07e091f3b545b449617e3a9d1862a8c5619d" exitCode=0 Dec 01 17:38:57 crc kubenswrapper[4868]: I1201 17:38:57.415592 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5clwr" Dec 01 17:38:57 crc kubenswrapper[4868]: I1201 17:38:57.415675 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5clwr" event={"ID":"25f64f3d-b811-4548-b12e-5d1bf8b1dc12","Type":"ContainerDied","Data":"75b57ec40d32c1a005276f349c3f07e091f3b545b449617e3a9d1862a8c5619d"} Dec 01 17:38:57 crc kubenswrapper[4868]: I1201 17:38:57.415793 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5clwr" event={"ID":"25f64f3d-b811-4548-b12e-5d1bf8b1dc12","Type":"ContainerDied","Data":"f3d20f04bee80fe66f5f34a6883b2a2a84525bdbaa69326c0ba6ca0a74af55f9"} Dec 01 17:38:57 crc kubenswrapper[4868]: I1201 17:38:57.415837 4868 scope.go:117] "RemoveContainer" containerID="75b57ec40d32c1a005276f349c3f07e091f3b545b449617e3a9d1862a8c5619d" Dec 01 17:38:57 crc kubenswrapper[4868]: I1201 17:38:57.445725 4868 scope.go:117] "RemoveContainer" containerID="75b57ec40d32c1a005276f349c3f07e091f3b545b449617e3a9d1862a8c5619d" Dec 01 17:38:57 crc kubenswrapper[4868]: E1201 17:38:57.447436 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75b57ec40d32c1a005276f349c3f07e091f3b545b449617e3a9d1862a8c5619d\": container with ID starting with 75b57ec40d32c1a005276f349c3f07e091f3b545b449617e3a9d1862a8c5619d not found: ID does not exist" containerID="75b57ec40d32c1a005276f349c3f07e091f3b545b449617e3a9d1862a8c5619d" Dec 01 17:38:57 crc kubenswrapper[4868]: I1201 17:38:57.447551 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75b57ec40d32c1a005276f349c3f07e091f3b545b449617e3a9d1862a8c5619d"} err="failed to get container status \"75b57ec40d32c1a005276f349c3f07e091f3b545b449617e3a9d1862a8c5619d\": rpc error: code = NotFound desc = could not find container \"75b57ec40d32c1a005276f349c3f07e091f3b545b449617e3a9d1862a8c5619d\": container with ID starting with 75b57ec40d32c1a005276f349c3f07e091f3b545b449617e3a9d1862a8c5619d not found: ID does not exist" Dec 01 17:38:57 crc kubenswrapper[4868]: I1201 17:38:57.467564 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-5clwr"] Dec 01 17:38:57 crc kubenswrapper[4868]: I1201 17:38:57.472271 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-5clwr"] Dec 01 17:38:58 crc kubenswrapper[4868]: I1201 17:38:58.185549 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25f64f3d-b811-4548-b12e-5d1bf8b1dc12" path="/var/lib/kubelet/pods/25f64f3d-b811-4548-b12e-5d1bf8b1dc12/volumes" Dec 01 17:39:05 crc kubenswrapper[4868]: I1201 17:39:05.331380 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-8rf2p" Dec 01 17:39:05 crc kubenswrapper[4868]: I1201 17:39:05.332146 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-8rf2p" Dec 01 17:39:05 crc kubenswrapper[4868]: I1201 17:39:05.358567 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-8rf2p" Dec 01 17:39:05 crc kubenswrapper[4868]: I1201 17:39:05.495923 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-8rf2p" Dec 01 17:39:06 crc kubenswrapper[4868]: I1201 17:39:06.593655 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-57x2t" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.039142 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts"] Dec 01 17:39:07 crc kubenswrapper[4868]: E1201 17:39:07.039454 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25f64f3d-b811-4548-b12e-5d1bf8b1dc12" containerName="registry-server" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.039734 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="25f64f3d-b811-4548-b12e-5d1bf8b1dc12" containerName="registry-server" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.040071 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="25f64f3d-b811-4548-b12e-5d1bf8b1dc12" containerName="registry-server" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.041093 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.044436 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-jjbmq" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.050104 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts"] Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.188791 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/127f8203-d151-4430-97ae-405c8788a2af-util\") pod \"3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts\" (UID: \"127f8203-d151-4430-97ae-405c8788a2af\") " pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.188838 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/127f8203-d151-4430-97ae-405c8788a2af-bundle\") pod \"3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts\" (UID: \"127f8203-d151-4430-97ae-405c8788a2af\") " pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.188911 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd548\" (UniqueName: \"kubernetes.io/projected/127f8203-d151-4430-97ae-405c8788a2af-kube-api-access-qd548\") pod \"3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts\" (UID: \"127f8203-d151-4430-97ae-405c8788a2af\") " pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.289928 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd548\" (UniqueName: \"kubernetes.io/projected/127f8203-d151-4430-97ae-405c8788a2af-kube-api-access-qd548\") pod \"3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts\" (UID: \"127f8203-d151-4430-97ae-405c8788a2af\") " pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.290538 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/127f8203-d151-4430-97ae-405c8788a2af-util\") pod \"3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts\" (UID: \"127f8203-d151-4430-97ae-405c8788a2af\") " pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.290796 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/127f8203-d151-4430-97ae-405c8788a2af-bundle\") pod \"3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts\" (UID: \"127f8203-d151-4430-97ae-405c8788a2af\") " pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.291217 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/127f8203-d151-4430-97ae-405c8788a2af-util\") pod \"3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts\" (UID: \"127f8203-d151-4430-97ae-405c8788a2af\") " pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.291237 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/127f8203-d151-4430-97ae-405c8788a2af-bundle\") pod \"3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts\" (UID: \"127f8203-d151-4430-97ae-405c8788a2af\") " pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.315273 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd548\" (UniqueName: \"kubernetes.io/projected/127f8203-d151-4430-97ae-405c8788a2af-kube-api-access-qd548\") pod \"3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts\" (UID: \"127f8203-d151-4430-97ae-405c8788a2af\") " pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.392342 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" Dec 01 17:39:07 crc kubenswrapper[4868]: I1201 17:39:07.607523 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts"] Dec 01 17:39:07 crc kubenswrapper[4868]: W1201 17:39:07.617192 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod127f8203_d151_4430_97ae_405c8788a2af.slice/crio-5046c65ec3eaf7ddad3a15ac50f26bf8f5fb453c77a62740f0a63d6c4f35235d WatchSource:0}: Error finding container 5046c65ec3eaf7ddad3a15ac50f26bf8f5fb453c77a62740f0a63d6c4f35235d: Status 404 returned error can't find the container with id 5046c65ec3eaf7ddad3a15ac50f26bf8f5fb453c77a62740f0a63d6c4f35235d Dec 01 17:39:08 crc kubenswrapper[4868]: I1201 17:39:08.493230 4868 generic.go:334] "Generic (PLEG): container finished" podID="127f8203-d151-4430-97ae-405c8788a2af" containerID="b68c9f32e40cecc16acf71b76178ce8c4da5d6be72af321627fd0ad169eca2d2" exitCode=0 Dec 01 17:39:08 crc kubenswrapper[4868]: I1201 17:39:08.493300 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" event={"ID":"127f8203-d151-4430-97ae-405c8788a2af","Type":"ContainerDied","Data":"b68c9f32e40cecc16acf71b76178ce8c4da5d6be72af321627fd0ad169eca2d2"} Dec 01 17:39:08 crc kubenswrapper[4868]: I1201 17:39:08.493370 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" event={"ID":"127f8203-d151-4430-97ae-405c8788a2af","Type":"ContainerStarted","Data":"5046c65ec3eaf7ddad3a15ac50f26bf8f5fb453c77a62740f0a63d6c4f35235d"} Dec 01 17:39:09 crc kubenswrapper[4868]: I1201 17:39:09.503740 4868 generic.go:334] "Generic (PLEG): container finished" podID="127f8203-d151-4430-97ae-405c8788a2af" containerID="381c536954021e7eee550e1ce600edcd4879140c14f3615bbd2819cd0f1da104" exitCode=0 Dec 01 17:39:09 crc kubenswrapper[4868]: I1201 17:39:09.503845 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" event={"ID":"127f8203-d151-4430-97ae-405c8788a2af","Type":"ContainerDied","Data":"381c536954021e7eee550e1ce600edcd4879140c14f3615bbd2819cd0f1da104"} Dec 01 17:39:10 crc kubenswrapper[4868]: I1201 17:39:10.514374 4868 generic.go:334] "Generic (PLEG): container finished" podID="127f8203-d151-4430-97ae-405c8788a2af" containerID="1a3adc08c58e87c64ff062d9b60540b782cc4eb30af799c3340e8b638995679d" exitCode=0 Dec 01 17:39:10 crc kubenswrapper[4868]: I1201 17:39:10.514438 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" event={"ID":"127f8203-d151-4430-97ae-405c8788a2af","Type":"ContainerDied","Data":"1a3adc08c58e87c64ff062d9b60540b782cc4eb30af799c3340e8b638995679d"} Dec 01 17:39:11 crc kubenswrapper[4868]: I1201 17:39:11.848383 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" Dec 01 17:39:11 crc kubenswrapper[4868]: I1201 17:39:11.966119 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd548\" (UniqueName: \"kubernetes.io/projected/127f8203-d151-4430-97ae-405c8788a2af-kube-api-access-qd548\") pod \"127f8203-d151-4430-97ae-405c8788a2af\" (UID: \"127f8203-d151-4430-97ae-405c8788a2af\") " Dec 01 17:39:11 crc kubenswrapper[4868]: I1201 17:39:11.966242 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/127f8203-d151-4430-97ae-405c8788a2af-util\") pod \"127f8203-d151-4430-97ae-405c8788a2af\" (UID: \"127f8203-d151-4430-97ae-405c8788a2af\") " Dec 01 17:39:11 crc kubenswrapper[4868]: I1201 17:39:11.966328 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/127f8203-d151-4430-97ae-405c8788a2af-bundle\") pod \"127f8203-d151-4430-97ae-405c8788a2af\" (UID: \"127f8203-d151-4430-97ae-405c8788a2af\") " Dec 01 17:39:11 crc kubenswrapper[4868]: I1201 17:39:11.967135 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/127f8203-d151-4430-97ae-405c8788a2af-bundle" (OuterVolumeSpecName: "bundle") pod "127f8203-d151-4430-97ae-405c8788a2af" (UID: "127f8203-d151-4430-97ae-405c8788a2af"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:39:11 crc kubenswrapper[4868]: I1201 17:39:11.974933 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/127f8203-d151-4430-97ae-405c8788a2af-kube-api-access-qd548" (OuterVolumeSpecName: "kube-api-access-qd548") pod "127f8203-d151-4430-97ae-405c8788a2af" (UID: "127f8203-d151-4430-97ae-405c8788a2af"). InnerVolumeSpecName "kube-api-access-qd548". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:39:12 crc kubenswrapper[4868]: I1201 17:39:12.001879 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/127f8203-d151-4430-97ae-405c8788a2af-util" (OuterVolumeSpecName: "util") pod "127f8203-d151-4430-97ae-405c8788a2af" (UID: "127f8203-d151-4430-97ae-405c8788a2af"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:39:12 crc kubenswrapper[4868]: I1201 17:39:12.068594 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd548\" (UniqueName: \"kubernetes.io/projected/127f8203-d151-4430-97ae-405c8788a2af-kube-api-access-qd548\") on node \"crc\" DevicePath \"\"" Dec 01 17:39:12 crc kubenswrapper[4868]: I1201 17:39:12.068671 4868 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/127f8203-d151-4430-97ae-405c8788a2af-util\") on node \"crc\" DevicePath \"\"" Dec 01 17:39:12 crc kubenswrapper[4868]: I1201 17:39:12.068699 4868 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/127f8203-d151-4430-97ae-405c8788a2af-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:39:12 crc kubenswrapper[4868]: I1201 17:39:12.535457 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" event={"ID":"127f8203-d151-4430-97ae-405c8788a2af","Type":"ContainerDied","Data":"5046c65ec3eaf7ddad3a15ac50f26bf8f5fb453c77a62740f0a63d6c4f35235d"} Dec 01 17:39:12 crc kubenswrapper[4868]: I1201 17:39:12.535998 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5046c65ec3eaf7ddad3a15ac50f26bf8f5fb453c77a62740f0a63d6c4f35235d" Dec 01 17:39:12 crc kubenswrapper[4868]: I1201 17:39:12.535519 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts" Dec 01 17:39:19 crc kubenswrapper[4868]: I1201 17:39:19.090767 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b559cfbd8-zw5pm"] Dec 01 17:39:19 crc kubenswrapper[4868]: E1201 17:39:19.095088 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127f8203-d151-4430-97ae-405c8788a2af" containerName="util" Dec 01 17:39:19 crc kubenswrapper[4868]: I1201 17:39:19.095182 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="127f8203-d151-4430-97ae-405c8788a2af" containerName="util" Dec 01 17:39:19 crc kubenswrapper[4868]: E1201 17:39:19.095242 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127f8203-d151-4430-97ae-405c8788a2af" containerName="extract" Dec 01 17:39:19 crc kubenswrapper[4868]: I1201 17:39:19.095337 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="127f8203-d151-4430-97ae-405c8788a2af" containerName="extract" Dec 01 17:39:19 crc kubenswrapper[4868]: E1201 17:39:19.095400 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127f8203-d151-4430-97ae-405c8788a2af" containerName="pull" Dec 01 17:39:19 crc kubenswrapper[4868]: I1201 17:39:19.095446 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="127f8203-d151-4430-97ae-405c8788a2af" containerName="pull" Dec 01 17:39:19 crc kubenswrapper[4868]: I1201 17:39:19.095620 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="127f8203-d151-4430-97ae-405c8788a2af" containerName="extract" Dec 01 17:39:19 crc kubenswrapper[4868]: I1201 17:39:19.096232 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7b559cfbd8-zw5pm" Dec 01 17:39:19 crc kubenswrapper[4868]: I1201 17:39:19.103729 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-cgvp7" Dec 01 17:39:19 crc kubenswrapper[4868]: I1201 17:39:19.116816 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b559cfbd8-zw5pm"] Dec 01 17:39:19 crc kubenswrapper[4868]: I1201 17:39:19.178572 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j7d5\" (UniqueName: \"kubernetes.io/projected/b6698ef2-c594-4859-ba6d-7d76127c819d-kube-api-access-4j7d5\") pod \"openstack-operator-controller-operator-7b559cfbd8-zw5pm\" (UID: \"b6698ef2-c594-4859-ba6d-7d76127c819d\") " pod="openstack-operators/openstack-operator-controller-operator-7b559cfbd8-zw5pm" Dec 01 17:39:19 crc kubenswrapper[4868]: I1201 17:39:19.279670 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j7d5\" (UniqueName: \"kubernetes.io/projected/b6698ef2-c594-4859-ba6d-7d76127c819d-kube-api-access-4j7d5\") pod \"openstack-operator-controller-operator-7b559cfbd8-zw5pm\" (UID: \"b6698ef2-c594-4859-ba6d-7d76127c819d\") " pod="openstack-operators/openstack-operator-controller-operator-7b559cfbd8-zw5pm" Dec 01 17:39:19 crc kubenswrapper[4868]: I1201 17:39:19.304909 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j7d5\" (UniqueName: \"kubernetes.io/projected/b6698ef2-c594-4859-ba6d-7d76127c819d-kube-api-access-4j7d5\") pod \"openstack-operator-controller-operator-7b559cfbd8-zw5pm\" (UID: \"b6698ef2-c594-4859-ba6d-7d76127c819d\") " pod="openstack-operators/openstack-operator-controller-operator-7b559cfbd8-zw5pm" Dec 01 17:39:19 crc kubenswrapper[4868]: I1201 17:39:19.419320 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7b559cfbd8-zw5pm" Dec 01 17:39:19 crc kubenswrapper[4868]: I1201 17:39:19.940624 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b559cfbd8-zw5pm"] Dec 01 17:39:20 crc kubenswrapper[4868]: I1201 17:39:20.622712 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b559cfbd8-zw5pm" event={"ID":"b6698ef2-c594-4859-ba6d-7d76127c819d","Type":"ContainerStarted","Data":"c2c3757bfa62cdb2d99558f5817f9a4b504a6f152cfaf7e676b41c73329c98cd"} Dec 01 17:39:25 crc kubenswrapper[4868]: I1201 17:39:25.904323 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:39:25 crc kubenswrapper[4868]: I1201 17:39:25.904659 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:39:25 crc kubenswrapper[4868]: I1201 17:39:25.904709 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:39:25 crc kubenswrapper[4868]: I1201 17:39:25.905577 4868 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5bcb6e04848aace81d63d47c22db36f1a32f6313299199f19533a292326b33e3"} pod="openshift-machine-config-operator/machine-config-daemon-njgpn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 17:39:25 crc kubenswrapper[4868]: I1201 17:39:25.905648 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" containerID="cri-o://5bcb6e04848aace81d63d47c22db36f1a32f6313299199f19533a292326b33e3" gracePeriod=600 Dec 01 17:39:26 crc kubenswrapper[4868]: I1201 17:39:26.691694 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b559cfbd8-zw5pm" event={"ID":"b6698ef2-c594-4859-ba6d-7d76127c819d","Type":"ContainerStarted","Data":"55f5ca9278a9c35fdd18d6f8ff56e55356281741d6252d600a0adf92137e489e"} Dec 01 17:39:26 crc kubenswrapper[4868]: I1201 17:39:26.692120 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7b559cfbd8-zw5pm" Dec 01 17:39:26 crc kubenswrapper[4868]: I1201 17:39:26.694461 4868 generic.go:334] "Generic (PLEG): container finished" podID="ec4e08bf-df56-4490-8e79-a4175c727405" containerID="5bcb6e04848aace81d63d47c22db36f1a32f6313299199f19533a292326b33e3" exitCode=0 Dec 01 17:39:26 crc kubenswrapper[4868]: I1201 17:39:26.694513 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerDied","Data":"5bcb6e04848aace81d63d47c22db36f1a32f6313299199f19533a292326b33e3"} Dec 01 17:39:26 crc kubenswrapper[4868]: I1201 17:39:26.694552 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"2d56a7ef7fc39bd11e02dfcafda9203c34f3593dd8178f744951427000e66a77"} Dec 01 17:39:26 crc kubenswrapper[4868]: I1201 17:39:26.694570 4868 scope.go:117] "RemoveContainer" containerID="af7782f09775c37fb2df7ea471d3f043178c78ecd6178cb2c2bc8c6ee7670877" Dec 01 17:39:26 crc kubenswrapper[4868]: I1201 17:39:26.734509 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7b559cfbd8-zw5pm" podStartSLOduration=1.660229208 podStartE2EDuration="7.73449075s" podCreationTimestamp="2025-12-01 17:39:19 +0000 UTC" firstStartedPulling="2025-12-01 17:39:19.964093105 +0000 UTC m=+832.335203516" lastFinishedPulling="2025-12-01 17:39:26.038354647 +0000 UTC m=+838.409465058" observedRunningTime="2025-12-01 17:39:26.72335812 +0000 UTC m=+839.094468551" watchObservedRunningTime="2025-12-01 17:39:26.73449075 +0000 UTC m=+839.105601161" Dec 01 17:39:39 crc kubenswrapper[4868]: I1201 17:39:39.424447 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7b559cfbd8-zw5pm" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.354889 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-8m6s9"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.356649 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8m6s9" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.359532 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-2mxr6" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.379571 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-pk4dk"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.381000 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pk4dk" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.383354 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-k5qzw" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.386247 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-4s44v"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.387565 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-4s44v" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.392634 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-f5nzv" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.403035 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-pk4dk"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.425166 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-rhj2q"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.426657 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-rhj2q" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.436203 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-mwgfz" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.457535 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-rhj2q"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.473184 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-8m6s9"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.494019 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-4s44v"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.495670 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-kbb5v"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.497354 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-kbb5v" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.501374 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-v8pfj" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.503563 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb9w8\" (UniqueName: \"kubernetes.io/projected/a11d06ed-a89a-4600-8c63-4a3845eb01e6-kube-api-access-kb9w8\") pod \"designate-operator-controller-manager-78b4bc895b-4s44v\" (UID: \"a11d06ed-a89a-4600-8c63-4a3845eb01e6\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-4s44v" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.503653 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2jdm\" (UniqueName: \"kubernetes.io/projected/f3eeed9b-dfbf-4e17-b80e-8792d471f1fb-kube-api-access-b2jdm\") pod \"barbican-operator-controller-manager-7d9dfd778-8m6s9\" (UID: \"f3eeed9b-dfbf-4e17-b80e-8792d471f1fb\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8m6s9" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.503701 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w28s8\" (UniqueName: \"kubernetes.io/projected/8cc7f247-f40c-4486-bd86-5f49036a4b2e-kube-api-access-w28s8\") pod \"cinder-operator-controller-manager-859b6ccc6-pk4dk\" (UID: \"8cc7f247-f40c-4486-bd86-5f49036a4b2e\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pk4dk" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.511013 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-kbb5v"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.514979 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5kkb2"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.516254 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5kkb2" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.544350 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-5rqsq" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.575033 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.577566 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.632095 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-668m4" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.637770 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lphst\" (UniqueName: \"kubernetes.io/projected/c5c836e6-8b19-44f8-91ae-4893dcbfd0fc-kube-api-access-lphst\") pod \"heat-operator-controller-manager-5f64f6f8bb-kbb5v\" (UID: \"c5c836e6-8b19-44f8-91ae-4893dcbfd0fc\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-kbb5v" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.637855 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb9w8\" (UniqueName: \"kubernetes.io/projected/a11d06ed-a89a-4600-8c63-4a3845eb01e6-kube-api-access-kb9w8\") pod \"designate-operator-controller-manager-78b4bc895b-4s44v\" (UID: \"a11d06ed-a89a-4600-8c63-4a3845eb01e6\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-4s44v" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.637910 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2jdm\" (UniqueName: \"kubernetes.io/projected/f3eeed9b-dfbf-4e17-b80e-8792d471f1fb-kube-api-access-b2jdm\") pod \"barbican-operator-controller-manager-7d9dfd778-8m6s9\" (UID: \"f3eeed9b-dfbf-4e17-b80e-8792d471f1fb\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8m6s9" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.637956 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w28s8\" (UniqueName: \"kubernetes.io/projected/8cc7f247-f40c-4486-bd86-5f49036a4b2e-kube-api-access-w28s8\") pod \"cinder-operator-controller-manager-859b6ccc6-pk4dk\" (UID: \"8cc7f247-f40c-4486-bd86-5f49036a4b2e\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pk4dk" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.638012 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp7g7\" (UniqueName: \"kubernetes.io/projected/6faacf07-6354-4314-829b-db239c85d98e-kube-api-access-tp7g7\") pod \"glance-operator-controller-manager-668d9c48b9-rhj2q\" (UID: \"6faacf07-6354-4314-829b-db239c85d98e\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-rhj2q" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.641483 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-c9nqt"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.641764 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.649056 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5kkb2"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.649185 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c9nqt" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.657982 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.661961 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-c9nqt"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.674428 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-xdvf2" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.681548 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb9w8\" (UniqueName: \"kubernetes.io/projected/a11d06ed-a89a-4600-8c63-4a3845eb01e6-kube-api-access-kb9w8\") pod \"designate-operator-controller-manager-78b4bc895b-4s44v\" (UID: \"a11d06ed-a89a-4600-8c63-4a3845eb01e6\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-4s44v" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.686248 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2jdm\" (UniqueName: \"kubernetes.io/projected/f3eeed9b-dfbf-4e17-b80e-8792d471f1fb-kube-api-access-b2jdm\") pod \"barbican-operator-controller-manager-7d9dfd778-8m6s9\" (UID: \"f3eeed9b-dfbf-4e17-b80e-8792d471f1fb\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8m6s9" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.691421 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w28s8\" (UniqueName: \"kubernetes.io/projected/8cc7f247-f40c-4486-bd86-5f49036a4b2e-kube-api-access-w28s8\") pod \"cinder-operator-controller-manager-859b6ccc6-pk4dk\" (UID: \"8cc7f247-f40c-4486-bd86-5f49036a4b2e\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pk4dk" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.702970 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pk4dk" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.702340 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-hqmbs"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.704994 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-hqmbs" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.711853 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-sn5vl" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.712507 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-4s44v" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.742911 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jfsv\" (UniqueName: \"kubernetes.io/projected/775bad96-f2da-4628-8242-af4778e37aee-kube-api-access-6jfsv\") pod \"infra-operator-controller-manager-55599fd5c4-2vg29\" (UID: \"775bad96-f2da-4628-8242-af4778e37aee\") " pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.743273 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert\") pod \"infra-operator-controller-manager-55599fd5c4-2vg29\" (UID: \"775bad96-f2da-4628-8242-af4778e37aee\") " pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.743410 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dztcl\" (UniqueName: \"kubernetes.io/projected/ccf55d21-7eac-4a3e-90c6-147287c4aac3-kube-api-access-dztcl\") pod \"keystone-operator-controller-manager-546d4bdf48-hqmbs\" (UID: \"ccf55d21-7eac-4a3e-90c6-147287c4aac3\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-hqmbs" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.743536 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp7g7\" (UniqueName: \"kubernetes.io/projected/6faacf07-6354-4314-829b-db239c85d98e-kube-api-access-tp7g7\") pod \"glance-operator-controller-manager-668d9c48b9-rhj2q\" (UID: \"6faacf07-6354-4314-829b-db239c85d98e\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-rhj2q" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.743971 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpj9t\" (UniqueName: \"kubernetes.io/projected/5cad124c-35fd-404e-bcf4-3b17e89e1288-kube-api-access-rpj9t\") pod \"ironic-operator-controller-manager-6c548fd776-c9nqt\" (UID: \"5cad124c-35fd-404e-bcf4-3b17e89e1288\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c9nqt" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.744153 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lphst\" (UniqueName: \"kubernetes.io/projected/c5c836e6-8b19-44f8-91ae-4893dcbfd0fc-kube-api-access-lphst\") pod \"heat-operator-controller-manager-5f64f6f8bb-kbb5v\" (UID: \"c5c836e6-8b19-44f8-91ae-4893dcbfd0fc\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-kbb5v" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.744534 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45wxf\" (UniqueName: \"kubernetes.io/projected/93fc4a4e-c6fb-4d43-88dd-0c24bcd85c5e-kube-api-access-45wxf\") pod \"horizon-operator-controller-manager-68c6d99b8f-5kkb2\" (UID: \"93fc4a4e-c6fb-4d43-88dd-0c24bcd85c5e\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5kkb2" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.748766 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-hqmbs"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.809067 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lphst\" (UniqueName: \"kubernetes.io/projected/c5c836e6-8b19-44f8-91ae-4893dcbfd0fc-kube-api-access-lphst\") pod \"heat-operator-controller-manager-5f64f6f8bb-kbb5v\" (UID: \"c5c836e6-8b19-44f8-91ae-4893dcbfd0fc\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-kbb5v" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.820043 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-jjvvf"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.821733 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-jjvvf" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.826500 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp7g7\" (UniqueName: \"kubernetes.io/projected/6faacf07-6354-4314-829b-db239c85d98e-kube-api-access-tp7g7\") pod \"glance-operator-controller-manager-668d9c48b9-rhj2q\" (UID: \"6faacf07-6354-4314-829b-db239c85d98e\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-rhj2q" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.826575 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-qknjx" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.833007 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9j2nw"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.836133 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9j2nw" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.839339 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-gg9d7" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.842624 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-kbb5v" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.845888 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jfsv\" (UniqueName: \"kubernetes.io/projected/775bad96-f2da-4628-8242-af4778e37aee-kube-api-access-6jfsv\") pod \"infra-operator-controller-manager-55599fd5c4-2vg29\" (UID: \"775bad96-f2da-4628-8242-af4778e37aee\") " pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.845928 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert\") pod \"infra-operator-controller-manager-55599fd5c4-2vg29\" (UID: \"775bad96-f2da-4628-8242-af4778e37aee\") " pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.845969 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dztcl\" (UniqueName: \"kubernetes.io/projected/ccf55d21-7eac-4a3e-90c6-147287c4aac3-kube-api-access-dztcl\") pod \"keystone-operator-controller-manager-546d4bdf48-hqmbs\" (UID: \"ccf55d21-7eac-4a3e-90c6-147287c4aac3\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-hqmbs" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.846012 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpj9t\" (UniqueName: \"kubernetes.io/projected/5cad124c-35fd-404e-bcf4-3b17e89e1288-kube-api-access-rpj9t\") pod \"ironic-operator-controller-manager-6c548fd776-c9nqt\" (UID: \"5cad124c-35fd-404e-bcf4-3b17e89e1288\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c9nqt" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.846048 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65l8c\" (UniqueName: \"kubernetes.io/projected/592806bc-2a42-4c69-b36f-1b8dc3c14249-kube-api-access-65l8c\") pod \"manila-operator-controller-manager-6546668bfd-jjvvf\" (UID: \"592806bc-2a42-4c69-b36f-1b8dc3c14249\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-jjvvf" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.846109 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sds25\" (UniqueName: \"kubernetes.io/projected/4e7cc027-32a6-437e-b05b-52c2984e0a61-kube-api-access-sds25\") pod \"mariadb-operator-controller-manager-56bbcc9d85-9j2nw\" (UID: \"4e7cc027-32a6-437e-b05b-52c2984e0a61\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9j2nw" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.846137 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45wxf\" (UniqueName: \"kubernetes.io/projected/93fc4a4e-c6fb-4d43-88dd-0c24bcd85c5e-kube-api-access-45wxf\") pod \"horizon-operator-controller-manager-68c6d99b8f-5kkb2\" (UID: \"93fc4a4e-c6fb-4d43-88dd-0c24bcd85c5e\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5kkb2" Dec 01 17:39:58 crc kubenswrapper[4868]: E1201 17:39:58.846721 4868 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 17:39:58 crc kubenswrapper[4868]: E1201 17:39:58.846852 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert podName:775bad96-f2da-4628-8242-af4778e37aee nodeName:}" failed. No retries permitted until 2025-12-01 17:39:59.346832012 +0000 UTC m=+871.717942423 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert") pod "infra-operator-controller-manager-55599fd5c4-2vg29" (UID: "775bad96-f2da-4628-8242-af4778e37aee") : secret "infra-operator-webhook-server-cert" not found Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.862182 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-jjvvf"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.882804 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45wxf\" (UniqueName: \"kubernetes.io/projected/93fc4a4e-c6fb-4d43-88dd-0c24bcd85c5e-kube-api-access-45wxf\") pod \"horizon-operator-controller-manager-68c6d99b8f-5kkb2\" (UID: \"93fc4a4e-c6fb-4d43-88dd-0c24bcd85c5e\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5kkb2" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.887673 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpj9t\" (UniqueName: \"kubernetes.io/projected/5cad124c-35fd-404e-bcf4-3b17e89e1288-kube-api-access-rpj9t\") pod \"ironic-operator-controller-manager-6c548fd776-c9nqt\" (UID: \"5cad124c-35fd-404e-bcf4-3b17e89e1288\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c9nqt" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.889623 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dztcl\" (UniqueName: \"kubernetes.io/projected/ccf55d21-7eac-4a3e-90c6-147287c4aac3-kube-api-access-dztcl\") pod \"keystone-operator-controller-manager-546d4bdf48-hqmbs\" (UID: \"ccf55d21-7eac-4a3e-90c6-147287c4aac3\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-hqmbs" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.894582 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jfsv\" (UniqueName: \"kubernetes.io/projected/775bad96-f2da-4628-8242-af4778e37aee-kube-api-access-6jfsv\") pod \"infra-operator-controller-manager-55599fd5c4-2vg29\" (UID: \"775bad96-f2da-4628-8242-af4778e37aee\") " pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.897002 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.898234 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.905357 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-j7qf8" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.909572 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.913509 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9j2nw"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.919453 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-f7k69"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.920682 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f7k69" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.928087 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-vhtlk" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.931132 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-f7k69"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.952097 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sds25\" (UniqueName: \"kubernetes.io/projected/4e7cc027-32a6-437e-b05b-52c2984e0a61-kube-api-access-sds25\") pod \"mariadb-operator-controller-manager-56bbcc9d85-9j2nw\" (UID: \"4e7cc027-32a6-437e-b05b-52c2984e0a61\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9j2nw" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.962447 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-fmkmw"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.965739 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fmkmw" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.967264 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c9nqt" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.968611 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-5k6w4" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.972058 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.972921 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65l8c\" (UniqueName: \"kubernetes.io/projected/592806bc-2a42-4c69-b36f-1b8dc3c14249-kube-api-access-65l8c\") pod \"manila-operator-controller-manager-6546668bfd-jjvvf\" (UID: \"592806bc-2a42-4c69-b36f-1b8dc3c14249\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-jjvvf" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.973499 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.979206 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.979247 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-5jrqr" Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.981585 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-fmkmw"] Dec 01 17:39:58 crc kubenswrapper[4868]: I1201 17:39:58.981851 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8m6s9" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.008126 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sds25\" (UniqueName: \"kubernetes.io/projected/4e7cc027-32a6-437e-b05b-52c2984e0a61-kube-api-access-sds25\") pod \"mariadb-operator-controller-manager-56bbcc9d85-9j2nw\" (UID: \"4e7cc027-32a6-437e-b05b-52c2984e0a61\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9j2nw" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.008642 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.033598 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65l8c\" (UniqueName: \"kubernetes.io/projected/592806bc-2a42-4c69-b36f-1b8dc3c14249-kube-api-access-65l8c\") pod \"manila-operator-controller-manager-6546668bfd-jjvvf\" (UID: \"592806bc-2a42-4c69-b36f-1b8dc3c14249\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-jjvvf" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.053073 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.054584 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.059344 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-pjt22" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.059922 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-rhj2q" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.074817 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-hqmbs" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.075407 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx6p6\" (UniqueName: \"kubernetes.io/projected/20c07d22-1628-40a4-97aa-605a2da611df-kube-api-access-bx6p6\") pod \"octavia-operator-controller-manager-998648c74-fmkmw\" (UID: \"20c07d22-1628-40a4-97aa-605a2da611df\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-fmkmw" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.075458 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5clsq\" (UniqueName: \"kubernetes.io/projected/b8df3e8b-0fc5-465f-b531-474518a1b809-kube-api-access-5clsq\") pod \"nova-operator-controller-manager-697bc559fc-f7k69\" (UID: \"b8df3e8b-0fc5-465f-b531-474518a1b809\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f7k69" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.075520 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ttbv\" (UniqueName: \"kubernetes.io/projected/d87f1b79-a4fb-4cfa-86d7-823ceedcf913-kube-api-access-2ttbv\") pod \"ovn-operator-controller-manager-b6456fdb6-97g6w\" (UID: \"d87f1b79-a4fb-4cfa-86d7-823ceedcf913\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.075560 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vb4m\" (UniqueName: \"kubernetes.io/projected/25f93b71-f929-4f74-a2d9-822f0b9402bb-kube-api-access-4vb4m\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs\" (UID: \"25f93b71-f929-4f74-a2d9-822f0b9402bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.075619 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs\" (UID: \"25f93b71-f929-4f74-a2d9-822f0b9402bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.075647 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdmvg\" (UniqueName: \"kubernetes.io/projected/97f09844-bdf4-4612-b4b4-966ae8dd49c2-kube-api-access-qdmvg\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-7mptb\" (UID: \"97f09844-bdf4-4612-b4b4-966ae8dd49c2\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.080006 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.081296 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.085019 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-zt6kt" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.089251 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.116636 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-g56bv"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.131657 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-g56bv" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.136370 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.146398 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-x5mtx" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.204012 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxn5s\" (UniqueName: \"kubernetes.io/projected/58e2eaed-6851-4568-a604-397a8944da7f-kube-api-access-cxn5s\") pod \"swift-operator-controller-manager-5f8c65bbfc-g56bv\" (UID: \"58e2eaed-6851-4568-a604-397a8944da7f\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-g56bv" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.204839 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ttbv\" (UniqueName: \"kubernetes.io/projected/d87f1b79-a4fb-4cfa-86d7-823ceedcf913-kube-api-access-2ttbv\") pod \"ovn-operator-controller-manager-b6456fdb6-97g6w\" (UID: \"d87f1b79-a4fb-4cfa-86d7-823ceedcf913\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.204970 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vb4m\" (UniqueName: \"kubernetes.io/projected/25f93b71-f929-4f74-a2d9-822f0b9402bb-kube-api-access-4vb4m\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs\" (UID: \"25f93b71-f929-4f74-a2d9-822f0b9402bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.205518 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs\" (UID: \"25f93b71-f929-4f74-a2d9-822f0b9402bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.205591 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdmvg\" (UniqueName: \"kubernetes.io/projected/97f09844-bdf4-4612-b4b4-966ae8dd49c2-kube-api-access-qdmvg\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-7mptb\" (UID: \"97f09844-bdf4-4612-b4b4-966ae8dd49c2\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.205854 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bbx2\" (UniqueName: \"kubernetes.io/projected/d5d35af7-2b67-42cd-ae7c-d73f105da292-kube-api-access-6bbx2\") pod \"placement-operator-controller-manager-78f8948974-mmjrx\" (UID: \"d5d35af7-2b67-42cd-ae7c-d73f105da292\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.206180 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx6p6\" (UniqueName: \"kubernetes.io/projected/20c07d22-1628-40a4-97aa-605a2da611df-kube-api-access-bx6p6\") pod \"octavia-operator-controller-manager-998648c74-fmkmw\" (UID: \"20c07d22-1628-40a4-97aa-605a2da611df\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-fmkmw" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.206243 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5clsq\" (UniqueName: \"kubernetes.io/projected/b8df3e8b-0fc5-465f-b531-474518a1b809-kube-api-access-5clsq\") pod \"nova-operator-controller-manager-697bc559fc-f7k69\" (UID: \"b8df3e8b-0fc5-465f-b531-474518a1b809\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f7k69" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.214037 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5kkb2" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.237025 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-jjvvf" Dec 01 17:39:59 crc kubenswrapper[4868]: E1201 17:39:59.238290 4868 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 17:39:59 crc kubenswrapper[4868]: E1201 17:39:59.238407 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert podName:25f93b71-f929-4f74-a2d9-822f0b9402bb nodeName:}" failed. No retries permitted until 2025-12-01 17:39:59.738353286 +0000 UTC m=+872.109463687 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" (UID: "25f93b71-f929-4f74-a2d9-822f0b9402bb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.244261 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9j2nw" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.245551 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.299637 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx6p6\" (UniqueName: \"kubernetes.io/projected/20c07d22-1628-40a4-97aa-605a2da611df-kube-api-access-bx6p6\") pod \"octavia-operator-controller-manager-998648c74-fmkmw\" (UID: \"20c07d22-1628-40a4-97aa-605a2da611df\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-fmkmw" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.301150 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ttbv\" (UniqueName: \"kubernetes.io/projected/d87f1b79-a4fb-4cfa-86d7-823ceedcf913-kube-api-access-2ttbv\") pod \"ovn-operator-controller-manager-b6456fdb6-97g6w\" (UID: \"d87f1b79-a4fb-4cfa-86d7-823ceedcf913\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.304599 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.309455 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bbx2\" (UniqueName: \"kubernetes.io/projected/d5d35af7-2b67-42cd-ae7c-d73f105da292-kube-api-access-6bbx2\") pod \"placement-operator-controller-manager-78f8948974-mmjrx\" (UID: \"d5d35af7-2b67-42cd-ae7c-d73f105da292\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.309530 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxn5s\" (UniqueName: \"kubernetes.io/projected/58e2eaed-6851-4568-a604-397a8944da7f-kube-api-access-cxn5s\") pod \"swift-operator-controller-manager-5f8c65bbfc-g56bv\" (UID: \"58e2eaed-6851-4568-a604-397a8944da7f\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-g56bv" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.309839 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vb4m\" (UniqueName: \"kubernetes.io/projected/25f93b71-f929-4f74-a2d9-822f0b9402bb-kube-api-access-4vb4m\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs\" (UID: \"25f93b71-f929-4f74-a2d9-822f0b9402bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.313295 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdmvg\" (UniqueName: \"kubernetes.io/projected/97f09844-bdf4-4612-b4b4-966ae8dd49c2-kube-api-access-qdmvg\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-7mptb\" (UID: \"97f09844-bdf4-4612-b4b4-966ae8dd49c2\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.318922 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-g56bv"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.320800 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-h9jlm" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.324773 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.326320 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.329312 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.337550 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-4tdlq" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.344653 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5clsq\" (UniqueName: \"kubernetes.io/projected/b8df3e8b-0fc5-465f-b531-474518a1b809-kube-api-access-5clsq\") pod \"nova-operator-controller-manager-697bc559fc-f7k69\" (UID: \"b8df3e8b-0fc5-465f-b531-474518a1b809\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f7k69" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.358393 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxn5s\" (UniqueName: \"kubernetes.io/projected/58e2eaed-6851-4568-a604-397a8944da7f-kube-api-access-cxn5s\") pod \"swift-operator-controller-manager-5f8c65bbfc-g56bv\" (UID: \"58e2eaed-6851-4568-a604-397a8944da7f\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-g56bv" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.365545 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-g56bv" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.366748 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bbx2\" (UniqueName: \"kubernetes.io/projected/d5d35af7-2b67-42cd-ae7c-d73f105da292-kube-api-access-6bbx2\") pod \"placement-operator-controller-manager-78f8948974-mmjrx\" (UID: \"d5d35af7-2b67-42cd-ae7c-d73f105da292\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.368352 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fmkmw" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.373056 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.415322 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48wtg\" (UniqueName: \"kubernetes.io/projected/8df8451f-54fc-4553-834b-37839d4c1807-kube-api-access-48wtg\") pod \"telemetry-operator-controller-manager-76cc84c6bb-l4hqj\" (UID: \"8df8451f-54fc-4553-834b-37839d4c1807\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.415397 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twbj9\" (UniqueName: \"kubernetes.io/projected/f1f47ec9-ea94-4fd5-b113-baa29c640369-kube-api-access-twbj9\") pod \"test-operator-controller-manager-5854674fcc-d2jx2\" (UID: \"f1f47ec9-ea94-4fd5-b113-baa29c640369\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.415491 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert\") pod \"infra-operator-controller-manager-55599fd5c4-2vg29\" (UID: \"775bad96-f2da-4628-8242-af4778e37aee\") " pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" Dec 01 17:39:59 crc kubenswrapper[4868]: E1201 17:39:59.415700 4868 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 17:39:59 crc kubenswrapper[4868]: E1201 17:39:59.415763 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert podName:775bad96-f2da-4628-8242-af4778e37aee nodeName:}" failed. No retries permitted until 2025-12-01 17:40:00.415742239 +0000 UTC m=+872.786852650 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert") pod "infra-operator-controller-manager-55599fd5c4-2vg29" (UID: "775bad96-f2da-4628-8242-af4778e37aee") : secret "infra-operator-webhook-server-cert" not found Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.422504 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.433456 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.439240 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-7vf5b" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.469642 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.482996 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.484592 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.492036 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-w7z49" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.492389 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.492505 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.519005 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dg9g\" (UniqueName: \"kubernetes.io/projected/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-kube-api-access-8dg9g\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.519094 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.519140 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58s8g\" (UniqueName: \"kubernetes.io/projected/fc4efb42-45ab-4341-b6f2-d95497223de3-kube-api-access-58s8g\") pod \"watcher-operator-controller-manager-769dc69bc-b78w5\" (UID: \"fc4efb42-45ab-4341-b6f2-d95497223de3\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.519174 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48wtg\" (UniqueName: \"kubernetes.io/projected/8df8451f-54fc-4553-834b-37839d4c1807-kube-api-access-48wtg\") pod \"telemetry-operator-controller-manager-76cc84c6bb-l4hqj\" (UID: \"8df8451f-54fc-4553-834b-37839d4c1807\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.519195 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twbj9\" (UniqueName: \"kubernetes.io/projected/f1f47ec9-ea94-4fd5-b113-baa29c640369-kube-api-access-twbj9\") pod \"test-operator-controller-manager-5854674fcc-d2jx2\" (UID: \"f1f47ec9-ea94-4fd5-b113-baa29c640369\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.519219 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.520161 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.524064 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.546993 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48wtg\" (UniqueName: \"kubernetes.io/projected/8df8451f-54fc-4553-834b-37839d4c1807-kube-api-access-48wtg\") pod \"telemetry-operator-controller-manager-76cc84c6bb-l4hqj\" (UID: \"8df8451f-54fc-4553-834b-37839d4c1807\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.566770 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twbj9\" (UniqueName: \"kubernetes.io/projected/f1f47ec9-ea94-4fd5-b113-baa29c640369-kube-api-access-twbj9\") pod \"test-operator-controller-manager-5854674fcc-d2jx2\" (UID: \"f1f47ec9-ea94-4fd5-b113-baa29c640369\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.570965 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-blw9f"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.572183 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-blw9f" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.579439 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-vqc7f" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.579818 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.600738 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-blw9f"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.628113 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f7k69" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.628165 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qf6g\" (UniqueName: \"kubernetes.io/projected/32e03cbc-4f29-49ee-8558-3f8950b0a383-kube-api-access-8qf6g\") pod \"rabbitmq-cluster-operator-manager-668c99d594-blw9f\" (UID: \"32e03cbc-4f29-49ee-8558-3f8950b0a383\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-blw9f" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.628234 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dg9g\" (UniqueName: \"kubernetes.io/projected/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-kube-api-access-8dg9g\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.628347 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.628387 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58s8g\" (UniqueName: \"kubernetes.io/projected/fc4efb42-45ab-4341-b6f2-d95497223de3-kube-api-access-58s8g\") pod \"watcher-operator-controller-manager-769dc69bc-b78w5\" (UID: \"fc4efb42-45ab-4341-b6f2-d95497223de3\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.628425 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:39:59 crc kubenswrapper[4868]: E1201 17:39:59.628571 4868 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 17:39:59 crc kubenswrapper[4868]: E1201 17:39:59.628655 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs podName:515c8a14-67ad-4c4c-8b3e-9433997fbdb3 nodeName:}" failed. No retries permitted until 2025-12-01 17:40:00.128632706 +0000 UTC m=+872.499743107 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs") pod "openstack-operator-controller-manager-755b486884-bvzjt" (UID: "515c8a14-67ad-4c4c-8b3e-9433997fbdb3") : secret "webhook-server-cert" not found Dec 01 17:39:59 crc kubenswrapper[4868]: E1201 17:39:59.628972 4868 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 17:39:59 crc kubenswrapper[4868]: E1201 17:39:59.629241 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs podName:515c8a14-67ad-4c4c-8b3e-9433997fbdb3 nodeName:}" failed. No retries permitted until 2025-12-01 17:40:00.129209783 +0000 UTC m=+872.500320354 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs") pod "openstack-operator-controller-manager-755b486884-bvzjt" (UID: "515c8a14-67ad-4c4c-8b3e-9433997fbdb3") : secret "metrics-server-cert" not found Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.651278 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58s8g\" (UniqueName: \"kubernetes.io/projected/fc4efb42-45ab-4341-b6f2-d95497223de3-kube-api-access-58s8g\") pod \"watcher-operator-controller-manager-769dc69bc-b78w5\" (UID: \"fc4efb42-45ab-4341-b6f2-d95497223de3\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.651327 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dg9g\" (UniqueName: \"kubernetes.io/projected/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-kube-api-access-8dg9g\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.660643 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.681642 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.713491 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.729215 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qf6g\" (UniqueName: \"kubernetes.io/projected/32e03cbc-4f29-49ee-8558-3f8950b0a383-kube-api-access-8qf6g\") pod \"rabbitmq-cluster-operator-manager-668c99d594-blw9f\" (UID: \"32e03cbc-4f29-49ee-8558-3f8950b0a383\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-blw9f" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.760103 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qf6g\" (UniqueName: \"kubernetes.io/projected/32e03cbc-4f29-49ee-8558-3f8950b0a383-kube-api-access-8qf6g\") pod \"rabbitmq-cluster-operator-manager-668c99d594-blw9f\" (UID: \"32e03cbc-4f29-49ee-8558-3f8950b0a383\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-blw9f" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.781172 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.831352 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-4s44v"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.845259 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs\" (UID: \"25f93b71-f929-4f74-a2d9-822f0b9402bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:39:59 crc kubenswrapper[4868]: E1201 17:39:59.845581 4868 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 17:39:59 crc kubenswrapper[4868]: E1201 17:39:59.845659 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert podName:25f93b71-f929-4f74-a2d9-822f0b9402bb nodeName:}" failed. No retries permitted until 2025-12-01 17:40:00.845634717 +0000 UTC m=+873.216745128 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" (UID: "25f93b71-f929-4f74-a2d9-822f0b9402bb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.847071 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-kbb5v"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.856681 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-pk4dk"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.921085 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-8m6s9"] Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.921996 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-blw9f" Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.978740 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-4s44v" event={"ID":"a11d06ed-a89a-4600-8c63-4a3845eb01e6","Type":"ContainerStarted","Data":"f1ca6a30417d3c55eddff08d72effb9a75cbc3dfef214157261e383cba5c36b4"} Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.983357 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pk4dk" event={"ID":"8cc7f247-f40c-4486-bd86-5f49036a4b2e","Type":"ContainerStarted","Data":"8ac7beae7869d56399e0bf222f99ece13ce8fc1ff32daec5c128413926658a30"} Dec 01 17:39:59 crc kubenswrapper[4868]: I1201 17:39:59.985724 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-kbb5v" event={"ID":"c5c836e6-8b19-44f8-91ae-4893dcbfd0fc","Type":"ContainerStarted","Data":"5d7dbab11020c22d56ae2dc20eec778b7d6b4232a7a2666009ba9cf81416e0ba"} Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.149030 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-hqmbs"] Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.155385 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-c9nqt"] Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.155798 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.155950 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.156014 4868 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.158306 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs podName:515c8a14-67ad-4c4c-8b3e-9433997fbdb3 nodeName:}" failed. No retries permitted until 2025-12-01 17:40:01.158268405 +0000 UTC m=+873.529378816 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs") pod "openstack-operator-controller-manager-755b486884-bvzjt" (UID: "515c8a14-67ad-4c4c-8b3e-9433997fbdb3") : secret "webhook-server-cert" not found Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.157281 4868 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.158509 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs podName:515c8a14-67ad-4c4c-8b3e-9433997fbdb3 nodeName:}" failed. No retries permitted until 2025-12-01 17:40:01.1584318 +0000 UTC m=+873.529542211 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs") pod "openstack-operator-controller-manager-755b486884-bvzjt" (UID: "515c8a14-67ad-4c4c-8b3e-9433997fbdb3") : secret "metrics-server-cert" not found Dec 01 17:40:00 crc kubenswrapper[4868]: W1201 17:40:00.181982 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cad124c_35fd_404e_bcf4_3b17e89e1288.slice/crio-2c2758447d945ad932a79435ed7fa4a9a4b09800d1126c88b2c41f179abeab7d WatchSource:0}: Error finding container 2c2758447d945ad932a79435ed7fa4a9a4b09800d1126c88b2c41f179abeab7d: Status 404 returned error can't find the container with id 2c2758447d945ad932a79435ed7fa4a9a4b09800d1126c88b2c41f179abeab7d Dec 01 17:40:00 crc kubenswrapper[4868]: W1201 17:40:00.190707 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccf55d21_7eac_4a3e_90c6_147287c4aac3.slice/crio-9ece8be2f2ed7028ada662fe567587597b6c1341df3d13129ca1615a3718616c WatchSource:0}: Error finding container 9ece8be2f2ed7028ada662fe567587597b6c1341df3d13129ca1615a3718616c: Status 404 returned error can't find the container with id 9ece8be2f2ed7028ada662fe567587597b6c1341df3d13129ca1615a3718616c Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.192127 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-rhj2q"] Dec 01 17:40:00 crc kubenswrapper[4868]: W1201 17:40:00.214615 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6faacf07_6354_4314_829b_db239c85d98e.slice/crio-e1c1b81b941d7c4c6bc351ff204c5bcc4161b05401b503787e3d6254896e3d77 WatchSource:0}: Error finding container e1c1b81b941d7c4c6bc351ff204c5bcc4161b05401b503787e3d6254896e3d77: Status 404 returned error can't find the container with id e1c1b81b941d7c4c6bc351ff204c5bcc4161b05401b503787e3d6254896e3d77 Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.250142 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9j2nw"] Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.254989 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-jjvvf"] Dec 01 17:40:00 crc kubenswrapper[4868]: W1201 17:40:00.274086 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod592806bc_2a42_4c69_b36f_1b8dc3c14249.slice/crio-5b56ca6a7d6bc4f2e04edf6a51e0e1841839d1636178bacc644cf62d68cd02cb WatchSource:0}: Error finding container 5b56ca6a7d6bc4f2e04edf6a51e0e1841839d1636178bacc644cf62d68cd02cb: Status 404 returned error can't find the container with id 5b56ca6a7d6bc4f2e04edf6a51e0e1841839d1636178bacc644cf62d68cd02cb Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.457660 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5kkb2"] Dec 01 17:40:00 crc kubenswrapper[4868]: W1201 17:40:00.459626 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93fc4a4e_c6fb_4d43_88dd_0c24bcd85c5e.slice/crio-177dbc2336c68cc15ba0d4b6a9a4e2ac1b1e160cec6305858801d4505b218bd4 WatchSource:0}: Error finding container 177dbc2336c68cc15ba0d4b6a9a4e2ac1b1e160cec6305858801d4505b218bd4: Status 404 returned error can't find the container with id 177dbc2336c68cc15ba0d4b6a9a4e2ac1b1e160cec6305858801d4505b218bd4 Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.471128 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert\") pod \"infra-operator-controller-manager-55599fd5c4-2vg29\" (UID: \"775bad96-f2da-4628-8242-af4778e37aee\") " pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.471353 4868 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.471416 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert podName:775bad96-f2da-4628-8242-af4778e37aee nodeName:}" failed. No retries permitted until 2025-12-01 17:40:02.471392396 +0000 UTC m=+874.842502807 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert") pod "infra-operator-controller-manager-55599fd5c4-2vg29" (UID: "775bad96-f2da-4628-8242-af4778e37aee") : secret "infra-operator-webhook-server-cert" not found Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.541597 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-f7k69"] Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.554656 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-g56bv"] Dec 01 17:40:00 crc kubenswrapper[4868]: W1201 17:40:00.556552 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58e2eaed_6851_4568_a604_397a8944da7f.slice/crio-9ff4d44e3064b4e4d140cc0c5ad85c9a4a31cebb528ef78b9ff69d770a33757c WatchSource:0}: Error finding container 9ff4d44e3064b4e4d140cc0c5ad85c9a4a31cebb528ef78b9ff69d770a33757c: Status 404 returned error can't find the container with id 9ff4d44e3064b4e4d140cc0c5ad85c9a4a31cebb528ef78b9ff69d770a33757c Dec 01 17:40:00 crc kubenswrapper[4868]: W1201 17:40:00.557205 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20c07d22_1628_40a4_97aa_605a2da611df.slice/crio-9deacec871645a035ebaec1179f3011ce28643fa1a7df43f7a1857547fd88e52 WatchSource:0}: Error finding container 9deacec871645a035ebaec1179f3011ce28643fa1a7df43f7a1857547fd88e52: Status 404 returned error can't find the container with id 9deacec871645a035ebaec1179f3011ce28643fa1a7df43f7a1857547fd88e52 Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.562599 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-fmkmw"] Dec 01 17:40:00 crc kubenswrapper[4868]: W1201 17:40:00.573614 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd87f1b79_a4fb_4cfa_86d7_823ceedcf913.slice/crio-be0816b4ccb7ae5ef75f60ed7e274d9b15f9f8a6bc3c0894bf38b7a9e2230234 WatchSource:0}: Error finding container be0816b4ccb7ae5ef75f60ed7e274d9b15f9f8a6bc3c0894bf38b7a9e2230234: Status 404 returned error can't find the container with id be0816b4ccb7ae5ef75f60ed7e274d9b15f9f8a6bc3c0894bf38b7a9e2230234 Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.573996 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w"] Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.576935 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2ttbv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-97g6w_openstack-operators(d87f1b79-a4fb-4cfa-86d7-823ceedcf913): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.578889 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2ttbv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-97g6w_openstack-operators(d87f1b79-a4fb-4cfa-86d7-823ceedcf913): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.580784 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w" podUID="d87f1b79-a4fb-4cfa-86d7-823ceedcf913" Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.659985 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj"] Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.667878 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-48wtg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-l4hqj_openstack-operators(8df8451f-54fc-4553-834b-37839d4c1807): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.675854 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-48wtg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-l4hqj_openstack-operators(8df8451f-54fc-4553-834b-37839d4c1807): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.677043 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj" podUID="8df8451f-54fc-4553-834b-37839d4c1807" Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.677150 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-twbj9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-d2jx2_openstack-operators(f1f47ec9-ea94-4fd5-b113-baa29c640369): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.677984 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2"] Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.679711 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-twbj9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-d2jx2_openstack-operators(f1f47ec9-ea94-4fd5-b113-baa29c640369): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.681031 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2" podUID="f1f47ec9-ea94-4fd5-b113-baa29c640369" Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.694088 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb"] Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.700190 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx"] Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.704414 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5"] Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.706859 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-58s8g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-b78w5_openstack-operators(fc4efb42-45ab-4341-b6f2-d95497223de3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 17:40:00 crc kubenswrapper[4868]: W1201 17:40:00.715338 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97f09844_bdf4_4612_b4b4_966ae8dd49c2.slice/crio-5bc2fcce4c48793556db316ee8ebbe15704d7f7dfe40db012ecf8ad710cd0285 WatchSource:0}: Error finding container 5bc2fcce4c48793556db316ee8ebbe15704d7f7dfe40db012ecf8ad710cd0285: Status 404 returned error can't find the container with id 5bc2fcce4c48793556db316ee8ebbe15704d7f7dfe40db012ecf8ad710cd0285 Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.718575 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-58s8g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-b78w5_openstack-operators(fc4efb42-45ab-4341-b6f2-d95497223de3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.720000 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5" podUID="fc4efb42-45ab-4341-b6f2-d95497223de3" Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.720276 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qdmvg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-7mptb_openstack-operators(97f09844-bdf4-4612-b4b4-966ae8dd49c2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.722373 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qdmvg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-7mptb_openstack-operators(97f09844-bdf4-4612-b4b4-966ae8dd49c2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.723548 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb" podUID="97f09844-bdf4-4612-b4b4-966ae8dd49c2" Dec 01 17:40:00 crc kubenswrapper[4868]: W1201 17:40:00.723809 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5d35af7_2b67_42cd_ae7c_d73f105da292.slice/crio-47cdd551f7c829a6ec57eb67bd69c88e60be60641771deaec1e28a94089f2f73 WatchSource:0}: Error finding container 47cdd551f7c829a6ec57eb67bd69c88e60be60641771deaec1e28a94089f2f73: Status 404 returned error can't find the container with id 47cdd551f7c829a6ec57eb67bd69c88e60be60641771deaec1e28a94089f2f73 Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.736447 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6bbx2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-mmjrx_openstack-operators(d5d35af7-2b67-42cd-ae7c-d73f105da292): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.739016 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6bbx2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-mmjrx_openstack-operators(d5d35af7-2b67-42cd-ae7c-d73f105da292): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.740239 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx" podUID="d5d35af7-2b67-42cd-ae7c-d73f105da292" Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.813068 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-blw9f"] Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.878588 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs\" (UID: \"25f93b71-f929-4f74-a2d9-822f0b9402bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.878837 4868 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 17:40:00 crc kubenswrapper[4868]: E1201 17:40:00.878958 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert podName:25f93b71-f929-4f74-a2d9-822f0b9402bb nodeName:}" failed. No retries permitted until 2025-12-01 17:40:02.87892195 +0000 UTC m=+875.250032351 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" (UID: "25f93b71-f929-4f74-a2d9-822f0b9402bb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.994510 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj" event={"ID":"8df8451f-54fc-4553-834b-37839d4c1807","Type":"ContainerStarted","Data":"ee7a9646ee2655568b6509d52bb1a39c1d7f7421b76abc5967eb8b0512a9564d"} Dec 01 17:40:00 crc kubenswrapper[4868]: I1201 17:40:00.996983 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb" event={"ID":"97f09844-bdf4-4612-b4b4-966ae8dd49c2","Type":"ContainerStarted","Data":"5bc2fcce4c48793556db316ee8ebbe15704d7f7dfe40db012ecf8ad710cd0285"} Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.002487 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-blw9f" event={"ID":"32e03cbc-4f29-49ee-8558-3f8950b0a383","Type":"ContainerStarted","Data":"c62638db7cc0298e624cb9c99501a7bf99148c05f4346eeac8b0f429d748652f"} Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.004447 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx" event={"ID":"d5d35af7-2b67-42cd-ae7c-d73f105da292","Type":"ContainerStarted","Data":"47cdd551f7c829a6ec57eb67bd69c88e60be60641771deaec1e28a94089f2f73"} Dec 01 17:40:01 crc kubenswrapper[4868]: E1201 17:40:01.005633 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb" podUID="97f09844-bdf4-4612-b4b4-966ae8dd49c2" Dec 01 17:40:01 crc kubenswrapper[4868]: E1201 17:40:01.006006 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj" podUID="8df8451f-54fc-4553-834b-37839d4c1807" Dec 01 17:40:01 crc kubenswrapper[4868]: E1201 17:40:01.011098 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx" podUID="d5d35af7-2b67-42cd-ae7c-d73f105da292" Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.020488 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5" event={"ID":"fc4efb42-45ab-4341-b6f2-d95497223de3","Type":"ContainerStarted","Data":"6d7daa963352a90e24da47ad81fa0732a9576435cdcae371978014939a7ab79a"} Dec 01 17:40:01 crc kubenswrapper[4868]: E1201 17:40:01.034024 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5" podUID="fc4efb42-45ab-4341-b6f2-d95497223de3" Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.035853 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8m6s9" event={"ID":"f3eeed9b-dfbf-4e17-b80e-8792d471f1fb","Type":"ContainerStarted","Data":"62038093bdfe555718db7fc9b3544bbe879e7f56ab0130f110619206fdeff009"} Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.038094 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-g56bv" event={"ID":"58e2eaed-6851-4568-a604-397a8944da7f","Type":"ContainerStarted","Data":"9ff4d44e3064b4e4d140cc0c5ad85c9a4a31cebb528ef78b9ff69d770a33757c"} Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.058156 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-jjvvf" event={"ID":"592806bc-2a42-4c69-b36f-1b8dc3c14249","Type":"ContainerStarted","Data":"5b56ca6a7d6bc4f2e04edf6a51e0e1841839d1636178bacc644cf62d68cd02cb"} Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.075035 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c9nqt" event={"ID":"5cad124c-35fd-404e-bcf4-3b17e89e1288","Type":"ContainerStarted","Data":"2c2758447d945ad932a79435ed7fa4a9a4b09800d1126c88b2c41f179abeab7d"} Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.080038 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f7k69" event={"ID":"b8df3e8b-0fc5-465f-b531-474518a1b809","Type":"ContainerStarted","Data":"c13a6e983d58984c4cf7b551649f714b9c146086467abce57193012277ef64e8"} Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.083008 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5kkb2" event={"ID":"93fc4a4e-c6fb-4d43-88dd-0c24bcd85c5e","Type":"ContainerStarted","Data":"177dbc2336c68cc15ba0d4b6a9a4e2ac1b1e160cec6305858801d4505b218bd4"} Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.089748 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2" event={"ID":"f1f47ec9-ea94-4fd5-b113-baa29c640369","Type":"ContainerStarted","Data":"2be5c50422ac38f9357fec2b6e7d9a9f80ea417bf089baff6d1f66a014944931"} Dec 01 17:40:01 crc kubenswrapper[4868]: E1201 17:40:01.091850 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2" podUID="f1f47ec9-ea94-4fd5-b113-baa29c640369" Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.092392 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-rhj2q" event={"ID":"6faacf07-6354-4314-829b-db239c85d98e","Type":"ContainerStarted","Data":"e1c1b81b941d7c4c6bc351ff204c5bcc4161b05401b503787e3d6254896e3d77"} Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.097129 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w" event={"ID":"d87f1b79-a4fb-4cfa-86d7-823ceedcf913","Type":"ContainerStarted","Data":"be0816b4ccb7ae5ef75f60ed7e274d9b15f9f8a6bc3c0894bf38b7a9e2230234"} Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.104444 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fmkmw" event={"ID":"20c07d22-1628-40a4-97aa-605a2da611df","Type":"ContainerStarted","Data":"9deacec871645a035ebaec1179f3011ce28643fa1a7df43f7a1857547fd88e52"} Dec 01 17:40:01 crc kubenswrapper[4868]: E1201 17:40:01.104781 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w" podUID="d87f1b79-a4fb-4cfa-86d7-823ceedcf913" Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.106468 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9j2nw" event={"ID":"4e7cc027-32a6-437e-b05b-52c2984e0a61","Type":"ContainerStarted","Data":"3a489b10ec6e6d97f28ca21020faa5c2d9cfca8622cf0d3f862b0f5c49fbc42d"} Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.108533 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-hqmbs" event={"ID":"ccf55d21-7eac-4a3e-90c6-147287c4aac3","Type":"ContainerStarted","Data":"9ece8be2f2ed7028ada662fe567587597b6c1341df3d13129ca1615a3718616c"} Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.187648 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:01 crc kubenswrapper[4868]: I1201 17:40:01.187783 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:01 crc kubenswrapper[4868]: E1201 17:40:01.188404 4868 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 17:40:01 crc kubenswrapper[4868]: E1201 17:40:01.188463 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs podName:515c8a14-67ad-4c4c-8b3e-9433997fbdb3 nodeName:}" failed. No retries permitted until 2025-12-01 17:40:03.188448792 +0000 UTC m=+875.559559203 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs") pod "openstack-operator-controller-manager-755b486884-bvzjt" (UID: "515c8a14-67ad-4c4c-8b3e-9433997fbdb3") : secret "webhook-server-cert" not found Dec 01 17:40:01 crc kubenswrapper[4868]: E1201 17:40:01.190091 4868 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 17:40:01 crc kubenswrapper[4868]: E1201 17:40:01.190178 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs podName:515c8a14-67ad-4c4c-8b3e-9433997fbdb3 nodeName:}" failed. No retries permitted until 2025-12-01 17:40:03.190159409 +0000 UTC m=+875.561269820 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs") pod "openstack-operator-controller-manager-755b486884-bvzjt" (UID: "515c8a14-67ad-4c4c-8b3e-9433997fbdb3") : secret "metrics-server-cert" not found Dec 01 17:40:02 crc kubenswrapper[4868]: E1201 17:40:02.150264 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx" podUID="d5d35af7-2b67-42cd-ae7c-d73f105da292" Dec 01 17:40:02 crc kubenswrapper[4868]: E1201 17:40:02.151226 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w" podUID="d87f1b79-a4fb-4cfa-86d7-823ceedcf913" Dec 01 17:40:02 crc kubenswrapper[4868]: E1201 17:40:02.151298 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5" podUID="fc4efb42-45ab-4341-b6f2-d95497223de3" Dec 01 17:40:02 crc kubenswrapper[4868]: E1201 17:40:02.151484 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj" podUID="8df8451f-54fc-4553-834b-37839d4c1807" Dec 01 17:40:02 crc kubenswrapper[4868]: E1201 17:40:02.154976 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb" podUID="97f09844-bdf4-4612-b4b4-966ae8dd49c2" Dec 01 17:40:02 crc kubenswrapper[4868]: E1201 17:40:02.175236 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2" podUID="f1f47ec9-ea94-4fd5-b113-baa29c640369" Dec 01 17:40:02 crc kubenswrapper[4868]: I1201 17:40:02.524998 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert\") pod \"infra-operator-controller-manager-55599fd5c4-2vg29\" (UID: \"775bad96-f2da-4628-8242-af4778e37aee\") " pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" Dec 01 17:40:02 crc kubenswrapper[4868]: E1201 17:40:02.525224 4868 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 17:40:02 crc kubenswrapper[4868]: E1201 17:40:02.525303 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert podName:775bad96-f2da-4628-8242-af4778e37aee nodeName:}" failed. No retries permitted until 2025-12-01 17:40:06.525282322 +0000 UTC m=+878.896392733 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert") pod "infra-operator-controller-manager-55599fd5c4-2vg29" (UID: "775bad96-f2da-4628-8242-af4778e37aee") : secret "infra-operator-webhook-server-cert" not found Dec 01 17:40:02 crc kubenswrapper[4868]: I1201 17:40:02.933839 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs\" (UID: \"25f93b71-f929-4f74-a2d9-822f0b9402bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:40:02 crc kubenswrapper[4868]: E1201 17:40:02.934074 4868 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 17:40:02 crc kubenswrapper[4868]: E1201 17:40:02.934190 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert podName:25f93b71-f929-4f74-a2d9-822f0b9402bb nodeName:}" failed. No retries permitted until 2025-12-01 17:40:06.934160733 +0000 UTC m=+879.305271144 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" (UID: "25f93b71-f929-4f74-a2d9-822f0b9402bb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 17:40:03 crc kubenswrapper[4868]: I1201 17:40:03.238908 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:03 crc kubenswrapper[4868]: I1201 17:40:03.239049 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:03 crc kubenswrapper[4868]: E1201 17:40:03.239207 4868 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 17:40:03 crc kubenswrapper[4868]: E1201 17:40:03.239274 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs podName:515c8a14-67ad-4c4c-8b3e-9433997fbdb3 nodeName:}" failed. No retries permitted until 2025-12-01 17:40:07.239255284 +0000 UTC m=+879.610365695 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs") pod "openstack-operator-controller-manager-755b486884-bvzjt" (UID: "515c8a14-67ad-4c4c-8b3e-9433997fbdb3") : secret "webhook-server-cert" not found Dec 01 17:40:03 crc kubenswrapper[4868]: E1201 17:40:03.239677 4868 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 17:40:03 crc kubenswrapper[4868]: E1201 17:40:03.239709 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs podName:515c8a14-67ad-4c4c-8b3e-9433997fbdb3 nodeName:}" failed. No retries permitted until 2025-12-01 17:40:07.239698386 +0000 UTC m=+879.610808797 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs") pod "openstack-operator-controller-manager-755b486884-bvzjt" (UID: "515c8a14-67ad-4c4c-8b3e-9433997fbdb3") : secret "metrics-server-cert" not found Dec 01 17:40:06 crc kubenswrapper[4868]: I1201 17:40:06.597215 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert\") pod \"infra-operator-controller-manager-55599fd5c4-2vg29\" (UID: \"775bad96-f2da-4628-8242-af4778e37aee\") " pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" Dec 01 17:40:06 crc kubenswrapper[4868]: E1201 17:40:06.597455 4868 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 17:40:06 crc kubenswrapper[4868]: E1201 17:40:06.597561 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert podName:775bad96-f2da-4628-8242-af4778e37aee nodeName:}" failed. No retries permitted until 2025-12-01 17:40:14.597536321 +0000 UTC m=+886.968646742 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert") pod "infra-operator-controller-manager-55599fd5c4-2vg29" (UID: "775bad96-f2da-4628-8242-af4778e37aee") : secret "infra-operator-webhook-server-cert" not found Dec 01 17:40:07 crc kubenswrapper[4868]: I1201 17:40:07.005772 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs\" (UID: \"25f93b71-f929-4f74-a2d9-822f0b9402bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:40:07 crc kubenswrapper[4868]: E1201 17:40:07.006509 4868 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 17:40:07 crc kubenswrapper[4868]: E1201 17:40:07.006596 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert podName:25f93b71-f929-4f74-a2d9-822f0b9402bb nodeName:}" failed. No retries permitted until 2025-12-01 17:40:15.006566996 +0000 UTC m=+887.377677407 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" (UID: "25f93b71-f929-4f74-a2d9-822f0b9402bb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 17:40:07 crc kubenswrapper[4868]: I1201 17:40:07.311546 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:07 crc kubenswrapper[4868]: E1201 17:40:07.311698 4868 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 17:40:07 crc kubenswrapper[4868]: E1201 17:40:07.311836 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs podName:515c8a14-67ad-4c4c-8b3e-9433997fbdb3 nodeName:}" failed. No retries permitted until 2025-12-01 17:40:15.311819801 +0000 UTC m=+887.682930212 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs") pod "openstack-operator-controller-manager-755b486884-bvzjt" (UID: "515c8a14-67ad-4c4c-8b3e-9433997fbdb3") : secret "webhook-server-cert" not found Dec 01 17:40:07 crc kubenswrapper[4868]: I1201 17:40:07.312228 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:07 crc kubenswrapper[4868]: E1201 17:40:07.312332 4868 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 17:40:07 crc kubenswrapper[4868]: E1201 17:40:07.312367 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs podName:515c8a14-67ad-4c4c-8b3e-9433997fbdb3 nodeName:}" failed. No retries permitted until 2025-12-01 17:40:15.312359345 +0000 UTC m=+887.683469746 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs") pod "openstack-operator-controller-manager-755b486884-bvzjt" (UID: "515c8a14-67ad-4c4c-8b3e-9433997fbdb3") : secret "metrics-server-cert" not found Dec 01 17:40:14 crc kubenswrapper[4868]: I1201 17:40:14.666572 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert\") pod \"infra-operator-controller-manager-55599fd5c4-2vg29\" (UID: \"775bad96-f2da-4628-8242-af4778e37aee\") " pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" Dec 01 17:40:14 crc kubenswrapper[4868]: I1201 17:40:14.675192 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/775bad96-f2da-4628-8242-af4778e37aee-cert\") pod \"infra-operator-controller-manager-55599fd5c4-2vg29\" (UID: \"775bad96-f2da-4628-8242-af4778e37aee\") " pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" Dec 01 17:40:14 crc kubenswrapper[4868]: I1201 17:40:14.848487 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" Dec 01 17:40:15 crc kubenswrapper[4868]: I1201 17:40:15.074363 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs\" (UID: \"25f93b71-f929-4f74-a2d9-822f0b9402bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:40:15 crc kubenswrapper[4868]: E1201 17:40:15.074592 4868 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 17:40:15 crc kubenswrapper[4868]: E1201 17:40:15.074680 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert podName:25f93b71-f929-4f74-a2d9-822f0b9402bb nodeName:}" failed. No retries permitted until 2025-12-01 17:40:31.074650583 +0000 UTC m=+903.445760994 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" (UID: "25f93b71-f929-4f74-a2d9-822f0b9402bb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 17:40:15 crc kubenswrapper[4868]: I1201 17:40:15.378834 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:15 crc kubenswrapper[4868]: I1201 17:40:15.379061 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:15 crc kubenswrapper[4868]: E1201 17:40:15.379232 4868 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 17:40:15 crc kubenswrapper[4868]: E1201 17:40:15.379264 4868 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 17:40:15 crc kubenswrapper[4868]: E1201 17:40:15.379286 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs podName:515c8a14-67ad-4c4c-8b3e-9433997fbdb3 nodeName:}" failed. No retries permitted until 2025-12-01 17:40:31.379270551 +0000 UTC m=+903.750380962 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs") pod "openstack-operator-controller-manager-755b486884-bvzjt" (UID: "515c8a14-67ad-4c4c-8b3e-9433997fbdb3") : secret "metrics-server-cert" not found Dec 01 17:40:15 crc kubenswrapper[4868]: E1201 17:40:15.379447 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs podName:515c8a14-67ad-4c4c-8b3e-9433997fbdb3 nodeName:}" failed. No retries permitted until 2025-12-01 17:40:31.379412065 +0000 UTC m=+903.750522636 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs") pod "openstack-operator-controller-manager-755b486884-bvzjt" (UID: "515c8a14-67ad-4c4c-8b3e-9433997fbdb3") : secret "webhook-server-cert" not found Dec 01 17:40:16 crc kubenswrapper[4868]: E1201 17:40:16.774498 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 01 17:40:16 crc kubenswrapper[4868]: E1201 17:40:16.775173 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8qf6g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-blw9f_openstack-operators(32e03cbc-4f29-49ee-8558-3f8950b0a383): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 17:40:16 crc kubenswrapper[4868]: E1201 17:40:16.776411 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-blw9f" podUID="32e03cbc-4f29-49ee-8558-3f8950b0a383" Dec 01 17:40:17 crc kubenswrapper[4868]: E1201 17:40:17.321872 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-blw9f" podUID="32e03cbc-4f29-49ee-8558-3f8950b0a383" Dec 01 17:40:17 crc kubenswrapper[4868]: E1201 17:40:17.433712 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3" Dec 01 17:40:17 crc kubenswrapper[4868]: E1201 17:40:17.433950 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dztcl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-546d4bdf48-hqmbs_openstack-operators(ccf55d21-7eac-4a3e-90c6-147287c4aac3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 17:40:18 crc kubenswrapper[4868]: E1201 17:40:18.130366 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 01 17:40:18 crc kubenswrapper[4868]: E1201 17:40:18.131143 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5clsq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-f7k69_openstack-operators(b8df3e8b-0fc5-465f-b531-474518a1b809): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 17:40:25 crc kubenswrapper[4868]: I1201 17:40:25.470880 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n7gp8"] Dec 01 17:40:25 crc kubenswrapper[4868]: I1201 17:40:25.473855 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:25 crc kubenswrapper[4868]: I1201 17:40:25.491614 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n7gp8"] Dec 01 17:40:25 crc kubenswrapper[4868]: I1201 17:40:25.582170 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f0c4cd5-91cc-4683-9bf4-38e725958049-catalog-content\") pod \"redhat-marketplace-n7gp8\" (UID: \"3f0c4cd5-91cc-4683-9bf4-38e725958049\") " pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:25 crc kubenswrapper[4868]: I1201 17:40:25.582231 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td8h4\" (UniqueName: \"kubernetes.io/projected/3f0c4cd5-91cc-4683-9bf4-38e725958049-kube-api-access-td8h4\") pod \"redhat-marketplace-n7gp8\" (UID: \"3f0c4cd5-91cc-4683-9bf4-38e725958049\") " pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:25 crc kubenswrapper[4868]: I1201 17:40:25.582346 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f0c4cd5-91cc-4683-9bf4-38e725958049-utilities\") pod \"redhat-marketplace-n7gp8\" (UID: \"3f0c4cd5-91cc-4683-9bf4-38e725958049\") " pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:25 crc kubenswrapper[4868]: I1201 17:40:25.683837 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f0c4cd5-91cc-4683-9bf4-38e725958049-catalog-content\") pod \"redhat-marketplace-n7gp8\" (UID: \"3f0c4cd5-91cc-4683-9bf4-38e725958049\") " pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:25 crc kubenswrapper[4868]: I1201 17:40:25.683895 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td8h4\" (UniqueName: \"kubernetes.io/projected/3f0c4cd5-91cc-4683-9bf4-38e725958049-kube-api-access-td8h4\") pod \"redhat-marketplace-n7gp8\" (UID: \"3f0c4cd5-91cc-4683-9bf4-38e725958049\") " pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:25 crc kubenswrapper[4868]: I1201 17:40:25.683985 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f0c4cd5-91cc-4683-9bf4-38e725958049-utilities\") pod \"redhat-marketplace-n7gp8\" (UID: \"3f0c4cd5-91cc-4683-9bf4-38e725958049\") " pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:25 crc kubenswrapper[4868]: I1201 17:40:25.684564 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f0c4cd5-91cc-4683-9bf4-38e725958049-utilities\") pod \"redhat-marketplace-n7gp8\" (UID: \"3f0c4cd5-91cc-4683-9bf4-38e725958049\") " pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:25 crc kubenswrapper[4868]: I1201 17:40:25.684568 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f0c4cd5-91cc-4683-9bf4-38e725958049-catalog-content\") pod \"redhat-marketplace-n7gp8\" (UID: \"3f0c4cd5-91cc-4683-9bf4-38e725958049\") " pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:25 crc kubenswrapper[4868]: I1201 17:40:25.713144 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td8h4\" (UniqueName: \"kubernetes.io/projected/3f0c4cd5-91cc-4683-9bf4-38e725958049-kube-api-access-td8h4\") pod \"redhat-marketplace-n7gp8\" (UID: \"3f0c4cd5-91cc-4683-9bf4-38e725958049\") " pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:25 crc kubenswrapper[4868]: I1201 17:40:25.804786 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:26 crc kubenswrapper[4868]: I1201 17:40:26.320725 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29"] Dec 01 17:40:26 crc kubenswrapper[4868]: I1201 17:40:26.389072 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pk4dk" event={"ID":"8cc7f247-f40c-4486-bd86-5f49036a4b2e","Type":"ContainerStarted","Data":"03b5fa6e210106bb0bc7bff2d2a03afa55a06a1e8b557dcfc2d34ba3c7f89e63"} Dec 01 17:40:27 crc kubenswrapper[4868]: W1201 17:40:27.368542 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod775bad96_f2da_4628_8242_af4778e37aee.slice/crio-85044871e7c0203f32f2588170dfc35de19f11109f2882c1f7ee9ffbc0665748 WatchSource:0}: Error finding container 85044871e7c0203f32f2588170dfc35de19f11109f2882c1f7ee9ffbc0665748: Status 404 returned error can't find the container with id 85044871e7c0203f32f2588170dfc35de19f11109f2882c1f7ee9ffbc0665748 Dec 01 17:40:27 crc kubenswrapper[4868]: I1201 17:40:27.402565 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" event={"ID":"775bad96-f2da-4628-8242-af4778e37aee","Type":"ContainerStarted","Data":"85044871e7c0203f32f2588170dfc35de19f11109f2882c1f7ee9ffbc0665748"} Dec 01 17:40:27 crc kubenswrapper[4868]: I1201 17:40:27.980211 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n7gp8"] Dec 01 17:40:28 crc kubenswrapper[4868]: W1201 17:40:28.086417 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f0c4cd5_91cc_4683_9bf4_38e725958049.slice/crio-74e4f884db79a1eb0d2c8f711b9a6815712e714054e33b3c148663a486aa63b4 WatchSource:0}: Error finding container 74e4f884db79a1eb0d2c8f711b9a6815712e714054e33b3c148663a486aa63b4: Status 404 returned error can't find the container with id 74e4f884db79a1eb0d2c8f711b9a6815712e714054e33b3c148663a486aa63b4 Dec 01 17:40:28 crc kubenswrapper[4868]: I1201 17:40:28.428552 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5kkb2" event={"ID":"93fc4a4e-c6fb-4d43-88dd-0c24bcd85c5e","Type":"ContainerStarted","Data":"459aea1c9b44b1133c8071ecb7aaef2999b0b035d2ada5f2ae94878081ebc75c"} Dec 01 17:40:28 crc kubenswrapper[4868]: I1201 17:40:28.449561 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8m6s9" event={"ID":"f3eeed9b-dfbf-4e17-b80e-8792d471f1fb","Type":"ContainerStarted","Data":"45bfeadf244f1104850c9fde735aa8fab9be59bdc3cb8a1e910412dd28b87b4c"} Dec 01 17:40:28 crc kubenswrapper[4868]: I1201 17:40:28.462666 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-kbb5v" event={"ID":"c5c836e6-8b19-44f8-91ae-4893dcbfd0fc","Type":"ContainerStarted","Data":"2d7456a41cdf0f016d1d093be650cf903413c42eb661d40b998280eedc793abd"} Dec 01 17:40:28 crc kubenswrapper[4868]: I1201 17:40:28.481315 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-g56bv" event={"ID":"58e2eaed-6851-4568-a604-397a8944da7f","Type":"ContainerStarted","Data":"a07f8826af7ea9cef3e47bc57ed80ff9adc3cda0a7b2c91c50e4905e5856cee8"} Dec 01 17:40:28 crc kubenswrapper[4868]: I1201 17:40:28.490488 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-jjvvf" event={"ID":"592806bc-2a42-4c69-b36f-1b8dc3c14249","Type":"ContainerStarted","Data":"0594c9aeb3e4e4d69244b0361d48bd7ff45c86d48f7c4e38e51bc491dca3402a"} Dec 01 17:40:28 crc kubenswrapper[4868]: I1201 17:40:28.493523 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj" event={"ID":"8df8451f-54fc-4553-834b-37839d4c1807","Type":"ContainerStarted","Data":"f469edf32e2aea75221db0ba6427fb8936b0cd420702c3465b085fd126d3a14e"} Dec 01 17:40:28 crc kubenswrapper[4868]: I1201 17:40:28.494721 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c9nqt" event={"ID":"5cad124c-35fd-404e-bcf4-3b17e89e1288","Type":"ContainerStarted","Data":"9fc0de3eac1503220845ccef35d6b8a1629d45a1616dacf63521a29053465b6e"} Dec 01 17:40:28 crc kubenswrapper[4868]: I1201 17:40:28.504449 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7gp8" event={"ID":"3f0c4cd5-91cc-4683-9bf4-38e725958049","Type":"ContainerStarted","Data":"74e4f884db79a1eb0d2c8f711b9a6815712e714054e33b3c148663a486aa63b4"} Dec 01 17:40:28 crc kubenswrapper[4868]: I1201 17:40:28.521999 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx" event={"ID":"d5d35af7-2b67-42cd-ae7c-d73f105da292","Type":"ContainerStarted","Data":"82ee600f54c3fc3b88b29dfaebb47c7feabd12249d05aab15a3f71fb31057444"} Dec 01 17:40:28 crc kubenswrapper[4868]: I1201 17:40:28.544326 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fmkmw" event={"ID":"20c07d22-1628-40a4-97aa-605a2da611df","Type":"ContainerStarted","Data":"88ded4a852c54d66186bc35575a5ce682dc1e88823f8be772ae062800ea4df68"} Dec 01 17:40:28 crc kubenswrapper[4868]: I1201 17:40:28.550658 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9j2nw" event={"ID":"4e7cc027-32a6-437e-b05b-52c2984e0a61","Type":"ContainerStarted","Data":"0f4944e54fbe8f75592ebaf92b60c024e4eab598f4846cf754988931c9d2ccc2"} Dec 01 17:40:28 crc kubenswrapper[4868]: I1201 17:40:28.566854 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-rhj2q" event={"ID":"6faacf07-6354-4314-829b-db239c85d98e","Type":"ContainerStarted","Data":"ec7e6159098ac72c1da6512b51e93ebc90ac8ca7cde6dcbf657c1b01b277a21f"} Dec 01 17:40:28 crc kubenswrapper[4868]: I1201 17:40:28.577016 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-4s44v" event={"ID":"a11d06ed-a89a-4600-8c63-4a3845eb01e6","Type":"ContainerStarted","Data":"f3181a709faee8a7c893bd8738e691f6e04caf16b41976f6597df2f1bce04f8f"} Dec 01 17:40:29 crc kubenswrapper[4868]: I1201 17:40:29.586835 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2" event={"ID":"f1f47ec9-ea94-4fd5-b113-baa29c640369","Type":"ContainerStarted","Data":"b2b5cc6941b9cfb76e3bc4c2ab5f071efa502909a191b734f43b7daa5f98d0f6"} Dec 01 17:40:29 crc kubenswrapper[4868]: I1201 17:40:29.590354 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb" event={"ID":"97f09844-bdf4-4612-b4b4-966ae8dd49c2","Type":"ContainerStarted","Data":"07f46eb19ff3f25e430a0e00e4fe3ec51428b4bfb471f68eabe191e86ecb60b1"} Dec 01 17:40:29 crc kubenswrapper[4868]: I1201 17:40:29.591669 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w" event={"ID":"d87f1b79-a4fb-4cfa-86d7-823ceedcf913","Type":"ContainerStarted","Data":"7bb980adeb99c5b36d5a1d651c7a3221f431bd5b1c855200249d2eac24144513"} Dec 01 17:40:30 crc kubenswrapper[4868]: I1201 17:40:30.619239 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5" event={"ID":"fc4efb42-45ab-4341-b6f2-d95497223de3","Type":"ContainerStarted","Data":"e637d585c3b8c521290f1547e3484facac78dd58d13ed747352b3837f06c9ab3"} Dec 01 17:40:30 crc kubenswrapper[4868]: I1201 17:40:30.621897 4868 generic.go:334] "Generic (PLEG): container finished" podID="3f0c4cd5-91cc-4683-9bf4-38e725958049" containerID="b33a0963b89e6f856f942806a8e3390f9242e947786b497279a1744dfe675f44" exitCode=0 Dec 01 17:40:30 crc kubenswrapper[4868]: I1201 17:40:30.621995 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7gp8" event={"ID":"3f0c4cd5-91cc-4683-9bf4-38e725958049","Type":"ContainerDied","Data":"b33a0963b89e6f856f942806a8e3390f9242e947786b497279a1744dfe675f44"} Dec 01 17:40:31 crc kubenswrapper[4868]: I1201 17:40:31.112804 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs\" (UID: \"25f93b71-f929-4f74-a2d9-822f0b9402bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:40:31 crc kubenswrapper[4868]: I1201 17:40:31.121693 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25f93b71-f929-4f74-a2d9-822f0b9402bb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs\" (UID: \"25f93b71-f929-4f74-a2d9-822f0b9402bb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:40:31 crc kubenswrapper[4868]: I1201 17:40:31.228373 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-5jrqr" Dec 01 17:40:31 crc kubenswrapper[4868]: I1201 17:40:31.234435 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:40:31 crc kubenswrapper[4868]: I1201 17:40:31.418568 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:31 crc kubenswrapper[4868]: I1201 17:40:31.418675 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:31 crc kubenswrapper[4868]: I1201 17:40:31.423891 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-metrics-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:31 crc kubenswrapper[4868]: I1201 17:40:31.427999 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/515c8a14-67ad-4c4c-8b3e-9433997fbdb3-webhook-certs\") pod \"openstack-operator-controller-manager-755b486884-bvzjt\" (UID: \"515c8a14-67ad-4c4c-8b3e-9433997fbdb3\") " pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:31 crc kubenswrapper[4868]: I1201 17:40:31.618418 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-w7z49" Dec 01 17:40:31 crc kubenswrapper[4868]: I1201 17:40:31.626387 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:32 crc kubenswrapper[4868]: I1201 17:40:32.638088 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs"] Dec 01 17:40:32 crc kubenswrapper[4868]: E1201 17:40:32.752666 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f7k69" podUID="b8df3e8b-0fc5-465f-b531-474518a1b809" Dec 01 17:40:32 crc kubenswrapper[4868]: I1201 17:40:32.954626 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt"] Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.692174 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-blw9f" event={"ID":"32e03cbc-4f29-49ee-8558-3f8950b0a383","Type":"ContainerStarted","Data":"5c29e1946168dfd757a4a8c92d69b632c539aae0f111e019eaaaed6c1a3d9cc3"} Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.724848 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pk4dk" event={"ID":"8cc7f247-f40c-4486-bd86-5f49036a4b2e","Type":"ContainerStarted","Data":"4b3e06772f3bfc3ae393a2715afc8c14d4acb8e54756861561080628dc869859"} Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.726193 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pk4dk" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.729483 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-blw9f" podStartSLOduration=5.247124188 podStartE2EDuration="34.729445229s" podCreationTimestamp="2025-12-01 17:39:59 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.8243388 +0000 UTC m=+873.195449211" lastFinishedPulling="2025-12-01 17:40:30.306659841 +0000 UTC m=+902.677770252" observedRunningTime="2025-12-01 17:40:33.718508699 +0000 UTC m=+906.089619110" watchObservedRunningTime="2025-12-01 17:40:33.729445229 +0000 UTC m=+906.100555660" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.732532 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pk4dk" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.748031 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-kbb5v" event={"ID":"c5c836e6-8b19-44f8-91ae-4893dcbfd0fc","Type":"ContainerStarted","Data":"e1c1b40a19cff248aa020f4cb074a9288700f06b472b3abecb9fcf79da9e25ac"} Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.748187 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-kbb5v" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.755960 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-kbb5v" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.767265 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-pk4dk" podStartSLOduration=3.327216694 podStartE2EDuration="35.767145465s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:39:59.959735622 +0000 UTC m=+872.330846033" lastFinishedPulling="2025-12-01 17:40:32.399664403 +0000 UTC m=+904.770774804" observedRunningTime="2025-12-01 17:40:33.766663022 +0000 UTC m=+906.137773433" watchObservedRunningTime="2025-12-01 17:40:33.767145465 +0000 UTC m=+906.138255876" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.771400 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-4s44v" event={"ID":"a11d06ed-a89a-4600-8c63-4a3845eb01e6","Type":"ContainerStarted","Data":"9aad19ee66c08081b565b1b412fef34a8dbce47cccd5799415c85ccfd37a2187"} Dec 01 17:40:33 crc kubenswrapper[4868]: E1201 17:40:33.771737 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-hqmbs" podUID="ccf55d21-7eac-4a3e-90c6-147287c4aac3" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.773454 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-4s44v" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.775470 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fmkmw" event={"ID":"20c07d22-1628-40a4-97aa-605a2da611df","Type":"ContainerStarted","Data":"4bad0cf02e4d3b543b1e622f206f504cfb296c84ebd694eebc9b6ef8ca4e5262"} Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.776273 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fmkmw" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.788641 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-4s44v" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.793832 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fmkmw" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.833705 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5" event={"ID":"fc4efb42-45ab-4341-b6f2-d95497223de3","Type":"ContainerStarted","Data":"eae81d5b7305eb9eb05cb7de77aa6e0af072b461efa52f6c09abefc33bbfd2e6"} Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.835588 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.839597 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" event={"ID":"25f93b71-f929-4f74-a2d9-822f0b9402bb","Type":"ContainerStarted","Data":"0a78dea621e12b2281afb478824c15e54de783b4aa7d31747a8c7e1d80325a12"} Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.841480 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-kbb5v" podStartSLOduration=3.246034904 podStartE2EDuration="35.841459646s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:39:59.960367689 +0000 UTC m=+872.331478100" lastFinishedPulling="2025-12-01 17:40:32.555792431 +0000 UTC m=+904.926902842" observedRunningTime="2025-12-01 17:40:33.789571261 +0000 UTC m=+906.160681682" watchObservedRunningTime="2025-12-01 17:40:33.841459646 +0000 UTC m=+906.212570067" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.847767 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-g56bv" event={"ID":"58e2eaed-6851-4568-a604-397a8944da7f","Type":"ContainerStarted","Data":"bfccda029a236cce5edced7aa0927edbc7403728a79e730e98d567508fdf172d"} Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.856473 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-g56bv" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.865017 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-g56bv" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.874448 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-4s44v" podStartSLOduration=3.414736489 podStartE2EDuration="35.874423742s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:39:59.943863066 +0000 UTC m=+872.314973477" lastFinishedPulling="2025-12-01 17:40:32.403550299 +0000 UTC m=+904.774660730" observedRunningTime="2025-12-01 17:40:33.867683837 +0000 UTC m=+906.238794248" watchObservedRunningTime="2025-12-01 17:40:33.874423742 +0000 UTC m=+906.245534153" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.887852 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c9nqt" event={"ID":"5cad124c-35fd-404e-bcf4-3b17e89e1288","Type":"ContainerStarted","Data":"5c4beba8213786ce8c48191e483359cf654f2771d3c4a773eb066bc66a436878"} Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.889136 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c9nqt" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.900262 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fmkmw" podStartSLOduration=4.038849882 podStartE2EDuration="35.900238721s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.560150794 +0000 UTC m=+872.931261205" lastFinishedPulling="2025-12-01 17:40:32.421539633 +0000 UTC m=+904.792650044" observedRunningTime="2025-12-01 17:40:33.898496813 +0000 UTC m=+906.269607224" watchObservedRunningTime="2025-12-01 17:40:33.900238721 +0000 UTC m=+906.271349132" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.908855 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c9nqt" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.910829 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb" event={"ID":"97f09844-bdf4-4612-b4b4-966ae8dd49c2","Type":"ContainerStarted","Data":"78bdb3881dccafcb1ac20537f77310de257a26683b1ebac9387374cf2fefc355"} Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.911562 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.933152 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx" event={"ID":"d5d35af7-2b67-42cd-ae7c-d73f105da292","Type":"ContainerStarted","Data":"8099a08b209a7832ab61fd670a26909f1a020748162a2b2d429a6ee067ff134e"} Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.934600 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.938483 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5" podStartSLOduration=3.210775674 podStartE2EDuration="34.938460061s" podCreationTimestamp="2025-12-01 17:39:59 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.706534855 +0000 UTC m=+873.077645266" lastFinishedPulling="2025-12-01 17:40:32.434219232 +0000 UTC m=+904.805329653" observedRunningTime="2025-12-01 17:40:33.934904553 +0000 UTC m=+906.306014964" watchObservedRunningTime="2025-12-01 17:40:33.938460061 +0000 UTC m=+906.309570462" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.948672 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.951524 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-jjvvf" event={"ID":"592806bc-2a42-4c69-b36f-1b8dc3c14249","Type":"ContainerStarted","Data":"f0dd05fcf6eb674e35bf1059799c88c6d024c0f630a4945b8501fa03f3c468fa"} Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.952334 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-jjvvf" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.970253 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-jjvvf" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.974434 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-g56bv" podStartSLOduration=4.121894433 podStartE2EDuration="35.974404308s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.560250867 +0000 UTC m=+872.931361278" lastFinishedPulling="2025-12-01 17:40:32.412760742 +0000 UTC m=+904.783871153" observedRunningTime="2025-12-01 17:40:33.96465745 +0000 UTC m=+906.335767861" watchObservedRunningTime="2025-12-01 17:40:33.974404308 +0000 UTC m=+906.345514719" Dec 01 17:40:33 crc kubenswrapper[4868]: I1201 17:40:33.975228 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f7k69" event={"ID":"b8df3e8b-0fc5-465f-b531-474518a1b809","Type":"ContainerStarted","Data":"5da2b7d5c0cbf09dab90ae5ac733247a248a439952e02ff9d980647b0e339d8e"} Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.039515 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-c9nqt" podStartSLOduration=3.831055224 podStartE2EDuration="36.039484616s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.198195592 +0000 UTC m=+872.569305993" lastFinishedPulling="2025-12-01 17:40:32.406624974 +0000 UTC m=+904.777735385" observedRunningTime="2025-12-01 17:40:34.001198144 +0000 UTC m=+906.372308555" watchObservedRunningTime="2025-12-01 17:40:34.039484616 +0000 UTC m=+906.410595027" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.066901 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb" podStartSLOduration=4.364513727 podStartE2EDuration="36.066876109s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.720025635 +0000 UTC m=+873.091136046" lastFinishedPulling="2025-12-01 17:40:32.422388017 +0000 UTC m=+904.793498428" observedRunningTime="2025-12-01 17:40:34.063435124 +0000 UTC m=+906.434545535" watchObservedRunningTime="2025-12-01 17:40:34.066876109 +0000 UTC m=+906.437986520" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.080368 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" event={"ID":"515c8a14-67ad-4c4c-8b3e-9433997fbdb3","Type":"ContainerStarted","Data":"dc2c4e9a9d72b89958ef97765176372238d25e527d67e041612176a689f6ef19"} Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.080508 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.080532 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.080565 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" event={"ID":"775bad96-f2da-4628-8242-af4778e37aee","Type":"ContainerStarted","Data":"1a7b7151d98e5932161139b02a9a1080e35c602b8885693ad1525fc5cb5f3efd"} Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.080586 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5kkb2" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.080598 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5kkb2" event={"ID":"93fc4a4e-c6fb-4d43-88dd-0c24bcd85c5e","Type":"ContainerStarted","Data":"094e7e7156fecc5bb5a10cf9f95ce8d58f01b38ad50ddf6fb404ceedfab16e35"} Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.084161 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9j2nw" event={"ID":"4e7cc027-32a6-437e-b05b-52c2984e0a61","Type":"ContainerStarted","Data":"1f4a566a03cbc246f07ee7d75c67ceb6f0d769f7a546b907531d9e35709afcd8"} Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.084246 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9j2nw" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.114252 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj" event={"ID":"8df8451f-54fc-4553-834b-37839d4c1807","Type":"ContainerStarted","Data":"1220b748664e548f0f20b1ade03f3e0d423f36af892fcfa3709777ed1834096a"} Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.130414 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.131611 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mmjrx" podStartSLOduration=4.603567334 podStartE2EDuration="36.131586996s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.736242711 +0000 UTC m=+873.107353122" lastFinishedPulling="2025-12-01 17:40:32.264262323 +0000 UTC m=+904.635372784" observedRunningTime="2025-12-01 17:40:34.100243855 +0000 UTC m=+906.471354266" watchObservedRunningTime="2025-12-01 17:40:34.131586996 +0000 UTC m=+906.502697397" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.143647 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5kkb2" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.143822 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2" event={"ID":"f1f47ec9-ea94-4fd5-b113-baa29c640369","Type":"ContainerStarted","Data":"e2450ca7e4a7e0e3003ea307f19191cd4ee194424cd70c950e3516b4e1048200"} Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.145058 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.152634 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9j2nw" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.156971 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5kkb2" podStartSLOduration=4.265171389 podStartE2EDuration="36.156923212s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.461553396 +0000 UTC m=+872.832663807" lastFinishedPulling="2025-12-01 17:40:32.353305209 +0000 UTC m=+904.724415630" observedRunningTime="2025-12-01 17:40:34.153031235 +0000 UTC m=+906.524141646" watchObservedRunningTime="2025-12-01 17:40:34.156923212 +0000 UTC m=+906.528033623" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.157490 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.158172 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.183758 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" podStartSLOduration=31.206794509 podStartE2EDuration="36.183728958s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:27.376207166 +0000 UTC m=+899.747317577" lastFinishedPulling="2025-12-01 17:40:32.353141615 +0000 UTC m=+904.724252026" observedRunningTime="2025-12-01 17:40:34.177168308 +0000 UTC m=+906.548278719" watchObservedRunningTime="2025-12-01 17:40:34.183728958 +0000 UTC m=+906.554839369" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.234751 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9j2nw" podStartSLOduration=4.093127213 podStartE2EDuration="36.234724419s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.273163571 +0000 UTC m=+872.644273982" lastFinishedPulling="2025-12-01 17:40:32.414760777 +0000 UTC m=+904.785871188" observedRunningTime="2025-12-01 17:40:34.225497155 +0000 UTC m=+906.596607576" watchObservedRunningTime="2025-12-01 17:40:34.234724419 +0000 UTC m=+906.605834830" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.325496 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" podStartSLOduration=35.325460741 podStartE2EDuration="35.325460741s" podCreationTimestamp="2025-12-01 17:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:40:34.31160034 +0000 UTC m=+906.682710751" watchObservedRunningTime="2025-12-01 17:40:34.325460741 +0000 UTC m=+906.696571152" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.340555 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-jjvvf" podStartSLOduration=4.194446806 podStartE2EDuration="36.340527365s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.279699821 +0000 UTC m=+872.650810222" lastFinishedPulling="2025-12-01 17:40:32.42578037 +0000 UTC m=+904.796890781" observedRunningTime="2025-12-01 17:40:34.336873765 +0000 UTC m=+906.707984176" watchObservedRunningTime="2025-12-01 17:40:34.340527365 +0000 UTC m=+906.711637766" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.400682 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-l4hqj" podStartSLOduration=4.714464699 podStartE2EDuration="36.399961588s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.667681477 +0000 UTC m=+873.038791888" lastFinishedPulling="2025-12-01 17:40:32.353178366 +0000 UTC m=+904.724288777" observedRunningTime="2025-12-01 17:40:34.386083867 +0000 UTC m=+906.757194288" watchObservedRunningTime="2025-12-01 17:40:34.399961588 +0000 UTC m=+906.771071999" Dec 01 17:40:34 crc kubenswrapper[4868]: I1201 17:40:34.459115 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-d2jx2" podStartSLOduration=3.7085158160000002 podStartE2EDuration="35.459097132s" podCreationTimestamp="2025-12-01 17:39:59 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.677011524 +0000 UTC m=+873.048121935" lastFinishedPulling="2025-12-01 17:40:32.42759282 +0000 UTC m=+904.798703251" observedRunningTime="2025-12-01 17:40:34.458671141 +0000 UTC m=+906.829781552" watchObservedRunningTime="2025-12-01 17:40:34.459097132 +0000 UTC m=+906.830207543" Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.155153 4868 generic.go:334] "Generic (PLEG): container finished" podID="3f0c4cd5-91cc-4683-9bf4-38e725958049" containerID="f682da9acdd7004061c5071ab0aa4e19f58db87989b657f262410acc96e9fce0" exitCode=0 Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.155288 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7gp8" event={"ID":"3f0c4cd5-91cc-4683-9bf4-38e725958049","Type":"ContainerDied","Data":"f682da9acdd7004061c5071ab0aa4e19f58db87989b657f262410acc96e9fce0"} Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.159196 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-rhj2q" event={"ID":"6faacf07-6354-4314-829b-db239c85d98e","Type":"ContainerStarted","Data":"b5e0964d6cbb0cfb69355b1af76063083ef1b5e0ede4f424c9388431cae8e221"} Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.160151 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-rhj2q" Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.162281 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-rhj2q" Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.163579 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" event={"ID":"515c8a14-67ad-4c4c-8b3e-9433997fbdb3","Type":"ContainerStarted","Data":"e67dcc079668173ff38c65af6c4d57132a6c14134af594881be78a7ff6d01e5f"} Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.166539 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w" event={"ID":"d87f1b79-a4fb-4cfa-86d7-823ceedcf913","Type":"ContainerStarted","Data":"3a8171bf64d08e15368cb40144f451c12110d4390d847d8d76542542e9a701d9"} Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.167285 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w" Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.169890 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w" Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.171832 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8m6s9" event={"ID":"f3eeed9b-dfbf-4e17-b80e-8792d471f1fb","Type":"ContainerStarted","Data":"a0013305bcef8e2eda6e59b926edcac4db702e3ed0c23556a908d856817ef79c"} Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.172241 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8m6s9" Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.173539 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-hqmbs" event={"ID":"ccf55d21-7eac-4a3e-90c6-147287c4aac3","Type":"ContainerStarted","Data":"87680339032e83baade44687e0f266552ef78eb658ff4231dfc39fee5a9e3e1a"} Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.175406 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8m6s9" Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.177062 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f7k69" event={"ID":"b8df3e8b-0fc5-465f-b531-474518a1b809","Type":"ContainerStarted","Data":"608a1776dbb6d71ddd48e5d8eb12cb1108417ca889d6091abd3426b01a0ad3e4"} Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.182165 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f7k69" Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.187803 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" event={"ID":"775bad96-f2da-4628-8242-af4778e37aee","Type":"ContainerStarted","Data":"d29cd325e5fcce95148530db26e1e2837e9cdf3815f62c481173b4d76c815854"} Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.222710 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-rhj2q" podStartSLOduration=5.018772738 podStartE2EDuration="37.222675866s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.217692987 +0000 UTC m=+872.588803398" lastFinishedPulling="2025-12-01 17:40:32.421596115 +0000 UTC m=+904.792706526" observedRunningTime="2025-12-01 17:40:35.205814773 +0000 UTC m=+907.576925184" watchObservedRunningTime="2025-12-01 17:40:35.222675866 +0000 UTC m=+907.593786277" Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.224197 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-7mptb" Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.224280 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-b78w5" Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.262454 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-97g6w" podStartSLOduration=5.366534132 podStartE2EDuration="37.262430149s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.576772361 +0000 UTC m=+872.947882772" lastFinishedPulling="2025-12-01 17:40:32.472668378 +0000 UTC m=+904.843778789" observedRunningTime="2025-12-01 17:40:35.257321208 +0000 UTC m=+907.628431609" watchObservedRunningTime="2025-12-01 17:40:35.262430149 +0000 UTC m=+907.633540570" Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.285811 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8m6s9" podStartSLOduration=4.8979412700000005 podStartE2EDuration="37.28579257s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.033716284 +0000 UTC m=+872.404826695" lastFinishedPulling="2025-12-01 17:40:32.421567564 +0000 UTC m=+904.792677995" observedRunningTime="2025-12-01 17:40:35.279299542 +0000 UTC m=+907.650409953" watchObservedRunningTime="2025-12-01 17:40:35.28579257 +0000 UTC m=+907.656902971" Dec 01 17:40:35 crc kubenswrapper[4868]: I1201 17:40:35.311080 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f7k69" podStartSLOduration=3.330063422 podStartE2EDuration="37.311058944s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.551428994 +0000 UTC m=+872.922539405" lastFinishedPulling="2025-12-01 17:40:34.532424516 +0000 UTC m=+906.903534927" observedRunningTime="2025-12-01 17:40:35.310648093 +0000 UTC m=+907.681758514" watchObservedRunningTime="2025-12-01 17:40:35.311058944 +0000 UTC m=+907.682169355" Dec 01 17:40:36 crc kubenswrapper[4868]: I1201 17:40:36.205277 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" event={"ID":"25f93b71-f929-4f74-a2d9-822f0b9402bb","Type":"ContainerStarted","Data":"ab086cd9baa8cbf59172fe9ae29e72fabd1053f338fab3bd52a1ebdde5349917"} Dec 01 17:40:36 crc kubenswrapper[4868]: I1201 17:40:36.232527 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-hqmbs" event={"ID":"ccf55d21-7eac-4a3e-90c6-147287c4aac3","Type":"ContainerStarted","Data":"15edb2224d767a95f64e114b83e76fe7c189aeb476b7093679f02f1c8a177172"} Dec 01 17:40:36 crc kubenswrapper[4868]: I1201 17:40:36.279204 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-hqmbs" podStartSLOduration=2.6291498779999998 podStartE2EDuration="38.279184366s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:00.197829012 +0000 UTC m=+872.568939423" lastFinishedPulling="2025-12-01 17:40:35.84786349 +0000 UTC m=+908.218973911" observedRunningTime="2025-12-01 17:40:36.270469887 +0000 UTC m=+908.641580298" watchObservedRunningTime="2025-12-01 17:40:36.279184366 +0000 UTC m=+908.650294777" Dec 01 17:40:37 crc kubenswrapper[4868]: I1201 17:40:37.242453 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7gp8" event={"ID":"3f0c4cd5-91cc-4683-9bf4-38e725958049","Type":"ContainerStarted","Data":"60d02cd775a696c94b7462cac0485d14f721e3f598750d90d4d357bd975de7ed"} Dec 01 17:40:37 crc kubenswrapper[4868]: I1201 17:40:37.246355 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" event={"ID":"25f93b71-f929-4f74-a2d9-822f0b9402bb","Type":"ContainerStarted","Data":"b8984884e86d419342d930b37bfdfca1417b067a31413047819810f04dd34c37"} Dec 01 17:40:37 crc kubenswrapper[4868]: I1201 17:40:37.246413 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:40:37 crc kubenswrapper[4868]: I1201 17:40:37.246847 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-hqmbs" Dec 01 17:40:37 crc kubenswrapper[4868]: I1201 17:40:37.268041 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n7gp8" podStartSLOduration=7.506002483 podStartE2EDuration="12.268020078s" podCreationTimestamp="2025-12-01 17:40:25 +0000 UTC" firstStartedPulling="2025-12-01 17:40:31.49306711 +0000 UTC m=+903.864177521" lastFinishedPulling="2025-12-01 17:40:36.255084705 +0000 UTC m=+908.626195116" observedRunningTime="2025-12-01 17:40:37.262741114 +0000 UTC m=+909.633851525" watchObservedRunningTime="2025-12-01 17:40:37.268020078 +0000 UTC m=+909.639130489" Dec 01 17:40:37 crc kubenswrapper[4868]: I1201 17:40:37.297368 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" podStartSLOduration=36.147351549 podStartE2EDuration="39.297334964s" podCreationTimestamp="2025-12-01 17:39:58 +0000 UTC" firstStartedPulling="2025-12-01 17:40:32.694786169 +0000 UTC m=+905.065896580" lastFinishedPulling="2025-12-01 17:40:35.844769584 +0000 UTC m=+908.215879995" observedRunningTime="2025-12-01 17:40:37.293423627 +0000 UTC m=+909.664534078" watchObservedRunningTime="2025-12-01 17:40:37.297334964 +0000 UTC m=+909.668445385" Dec 01 17:40:39 crc kubenswrapper[4868]: I1201 17:40:39.632681 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f7k69" Dec 01 17:40:39 crc kubenswrapper[4868]: I1201 17:40:39.763666 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rjfhx"] Dec 01 17:40:39 crc kubenswrapper[4868]: I1201 17:40:39.766911 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:39 crc kubenswrapper[4868]: I1201 17:40:39.781762 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rjfhx"] Dec 01 17:40:39 crc kubenswrapper[4868]: I1201 17:40:39.799909 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sspb2\" (UniqueName: \"kubernetes.io/projected/e61d330e-2486-4ed9-84a5-6600251afb44-kube-api-access-sspb2\") pod \"community-operators-rjfhx\" (UID: \"e61d330e-2486-4ed9-84a5-6600251afb44\") " pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:39 crc kubenswrapper[4868]: I1201 17:40:39.800367 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e61d330e-2486-4ed9-84a5-6600251afb44-catalog-content\") pod \"community-operators-rjfhx\" (UID: \"e61d330e-2486-4ed9-84a5-6600251afb44\") " pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:39 crc kubenswrapper[4868]: I1201 17:40:39.800527 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e61d330e-2486-4ed9-84a5-6600251afb44-utilities\") pod \"community-operators-rjfhx\" (UID: \"e61d330e-2486-4ed9-84a5-6600251afb44\") " pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:39 crc kubenswrapper[4868]: I1201 17:40:39.902542 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e61d330e-2486-4ed9-84a5-6600251afb44-catalog-content\") pod \"community-operators-rjfhx\" (UID: \"e61d330e-2486-4ed9-84a5-6600251afb44\") " pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:39 crc kubenswrapper[4868]: I1201 17:40:39.902628 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e61d330e-2486-4ed9-84a5-6600251afb44-utilities\") pod \"community-operators-rjfhx\" (UID: \"e61d330e-2486-4ed9-84a5-6600251afb44\") " pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:39 crc kubenswrapper[4868]: I1201 17:40:39.902674 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sspb2\" (UniqueName: \"kubernetes.io/projected/e61d330e-2486-4ed9-84a5-6600251afb44-kube-api-access-sspb2\") pod \"community-operators-rjfhx\" (UID: \"e61d330e-2486-4ed9-84a5-6600251afb44\") " pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:39 crc kubenswrapper[4868]: I1201 17:40:39.903296 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e61d330e-2486-4ed9-84a5-6600251afb44-catalog-content\") pod \"community-operators-rjfhx\" (UID: \"e61d330e-2486-4ed9-84a5-6600251afb44\") " pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:39 crc kubenswrapper[4868]: I1201 17:40:39.903470 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e61d330e-2486-4ed9-84a5-6600251afb44-utilities\") pod \"community-operators-rjfhx\" (UID: \"e61d330e-2486-4ed9-84a5-6600251afb44\") " pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:39 crc kubenswrapper[4868]: I1201 17:40:39.935456 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sspb2\" (UniqueName: \"kubernetes.io/projected/e61d330e-2486-4ed9-84a5-6600251afb44-kube-api-access-sspb2\") pod \"community-operators-rjfhx\" (UID: \"e61d330e-2486-4ed9-84a5-6600251afb44\") " pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:40 crc kubenswrapper[4868]: I1201 17:40:40.091591 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:40 crc kubenswrapper[4868]: I1201 17:40:40.658148 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rjfhx"] Dec 01 17:40:41 crc kubenswrapper[4868]: I1201 17:40:41.242988 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs" Dec 01 17:40:41 crc kubenswrapper[4868]: I1201 17:40:41.303885 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjfhx" event={"ID":"e61d330e-2486-4ed9-84a5-6600251afb44","Type":"ContainerStarted","Data":"53bbd2de1897a49cc884e038697634eae25eedd0006c8e3f9867cf166d9379c8"} Dec 01 17:40:41 crc kubenswrapper[4868]: I1201 17:40:41.637338 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-755b486884-bvzjt" Dec 01 17:40:42 crc kubenswrapper[4868]: I1201 17:40:42.319926 4868 generic.go:334] "Generic (PLEG): container finished" podID="e61d330e-2486-4ed9-84a5-6600251afb44" containerID="0733de993a4eb0f5cf01faf6b52fd2d05c3a3a749bdcb1a5f3677da9e77cd786" exitCode=0 Dec 01 17:40:42 crc kubenswrapper[4868]: I1201 17:40:42.320027 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjfhx" event={"ID":"e61d330e-2486-4ed9-84a5-6600251afb44","Type":"ContainerDied","Data":"0733de993a4eb0f5cf01faf6b52fd2d05c3a3a749bdcb1a5f3677da9e77cd786"} Dec 01 17:40:44 crc kubenswrapper[4868]: I1201 17:40:44.339144 4868 generic.go:334] "Generic (PLEG): container finished" podID="e61d330e-2486-4ed9-84a5-6600251afb44" containerID="90a4877ee7423aface54c4fa40a8874a7153c75fb20180ea42618faeac6a59c2" exitCode=0 Dec 01 17:40:44 crc kubenswrapper[4868]: I1201 17:40:44.339852 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjfhx" event={"ID":"e61d330e-2486-4ed9-84a5-6600251afb44","Type":"ContainerDied","Data":"90a4877ee7423aface54c4fa40a8874a7153c75fb20180ea42618faeac6a59c2"} Dec 01 17:40:44 crc kubenswrapper[4868]: I1201 17:40:44.857047 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-55599fd5c4-2vg29" Dec 01 17:40:45 crc kubenswrapper[4868]: I1201 17:40:45.357021 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjfhx" event={"ID":"e61d330e-2486-4ed9-84a5-6600251afb44","Type":"ContainerStarted","Data":"5d8680ee0f53548ea9c7df11ded7b08b91dc376772db5165e8b5df52dcf5cbc2"} Dec 01 17:40:45 crc kubenswrapper[4868]: I1201 17:40:45.378576 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rjfhx" podStartSLOduration=3.526264485 podStartE2EDuration="6.378556332s" podCreationTimestamp="2025-12-01 17:40:39 +0000 UTC" firstStartedPulling="2025-12-01 17:40:42.323374291 +0000 UTC m=+914.694484712" lastFinishedPulling="2025-12-01 17:40:45.175666138 +0000 UTC m=+917.546776559" observedRunningTime="2025-12-01 17:40:45.374910431 +0000 UTC m=+917.746020842" watchObservedRunningTime="2025-12-01 17:40:45.378556332 +0000 UTC m=+917.749666743" Dec 01 17:40:45 crc kubenswrapper[4868]: I1201 17:40:45.805860 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:45 crc kubenswrapper[4868]: I1201 17:40:45.805902 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:45 crc kubenswrapper[4868]: I1201 17:40:45.865711 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:46 crc kubenswrapper[4868]: I1201 17:40:46.415544 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:47 crc kubenswrapper[4868]: I1201 17:40:47.125798 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n7gp8"] Dec 01 17:40:48 crc kubenswrapper[4868]: I1201 17:40:48.384013 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n7gp8" podUID="3f0c4cd5-91cc-4683-9bf4-38e725958049" containerName="registry-server" containerID="cri-o://60d02cd775a696c94b7462cac0485d14f721e3f598750d90d4d357bd975de7ed" gracePeriod=2 Dec 01 17:40:49 crc kubenswrapper[4868]: I1201 17:40:49.079127 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-hqmbs" Dec 01 17:40:49 crc kubenswrapper[4868]: I1201 17:40:49.396194 4868 generic.go:334] "Generic (PLEG): container finished" podID="3f0c4cd5-91cc-4683-9bf4-38e725958049" containerID="60d02cd775a696c94b7462cac0485d14f721e3f598750d90d4d357bd975de7ed" exitCode=0 Dec 01 17:40:49 crc kubenswrapper[4868]: I1201 17:40:49.396261 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7gp8" event={"ID":"3f0c4cd5-91cc-4683-9bf4-38e725958049","Type":"ContainerDied","Data":"60d02cd775a696c94b7462cac0485d14f721e3f598750d90d4d357bd975de7ed"} Dec 01 17:40:50 crc kubenswrapper[4868]: I1201 17:40:50.092282 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:50 crc kubenswrapper[4868]: I1201 17:40:50.092342 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:50 crc kubenswrapper[4868]: I1201 17:40:50.147373 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:50 crc kubenswrapper[4868]: I1201 17:40:50.450801 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:50 crc kubenswrapper[4868]: I1201 17:40:50.524340 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rjfhx"] Dec 01 17:40:50 crc kubenswrapper[4868]: I1201 17:40:50.851775 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:50 crc kubenswrapper[4868]: I1201 17:40:50.997012 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f0c4cd5-91cc-4683-9bf4-38e725958049-catalog-content\") pod \"3f0c4cd5-91cc-4683-9bf4-38e725958049\" (UID: \"3f0c4cd5-91cc-4683-9bf4-38e725958049\") " Dec 01 17:40:50 crc kubenswrapper[4868]: I1201 17:40:50.997357 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-td8h4\" (UniqueName: \"kubernetes.io/projected/3f0c4cd5-91cc-4683-9bf4-38e725958049-kube-api-access-td8h4\") pod \"3f0c4cd5-91cc-4683-9bf4-38e725958049\" (UID: \"3f0c4cd5-91cc-4683-9bf4-38e725958049\") " Dec 01 17:40:50 crc kubenswrapper[4868]: I1201 17:40:50.997403 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f0c4cd5-91cc-4683-9bf4-38e725958049-utilities\") pod \"3f0c4cd5-91cc-4683-9bf4-38e725958049\" (UID: \"3f0c4cd5-91cc-4683-9bf4-38e725958049\") " Dec 01 17:40:50 crc kubenswrapper[4868]: I1201 17:40:50.998260 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f0c4cd5-91cc-4683-9bf4-38e725958049-utilities" (OuterVolumeSpecName: "utilities") pod "3f0c4cd5-91cc-4683-9bf4-38e725958049" (UID: "3f0c4cd5-91cc-4683-9bf4-38e725958049"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:40:51 crc kubenswrapper[4868]: I1201 17:40:51.004063 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f0c4cd5-91cc-4683-9bf4-38e725958049-kube-api-access-td8h4" (OuterVolumeSpecName: "kube-api-access-td8h4") pod "3f0c4cd5-91cc-4683-9bf4-38e725958049" (UID: "3f0c4cd5-91cc-4683-9bf4-38e725958049"). InnerVolumeSpecName "kube-api-access-td8h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:40:51 crc kubenswrapper[4868]: I1201 17:40:51.015553 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f0c4cd5-91cc-4683-9bf4-38e725958049-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f0c4cd5-91cc-4683-9bf4-38e725958049" (UID: "3f0c4cd5-91cc-4683-9bf4-38e725958049"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:40:51 crc kubenswrapper[4868]: I1201 17:40:51.100353 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-td8h4\" (UniqueName: \"kubernetes.io/projected/3f0c4cd5-91cc-4683-9bf4-38e725958049-kube-api-access-td8h4\") on node \"crc\" DevicePath \"\"" Dec 01 17:40:51 crc kubenswrapper[4868]: I1201 17:40:51.100408 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f0c4cd5-91cc-4683-9bf4-38e725958049-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:40:51 crc kubenswrapper[4868]: I1201 17:40:51.100424 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f0c4cd5-91cc-4683-9bf4-38e725958049-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:40:51 crc kubenswrapper[4868]: I1201 17:40:51.414993 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7gp8" event={"ID":"3f0c4cd5-91cc-4683-9bf4-38e725958049","Type":"ContainerDied","Data":"74e4f884db79a1eb0d2c8f711b9a6815712e714054e33b3c148663a486aa63b4"} Dec 01 17:40:51 crc kubenswrapper[4868]: I1201 17:40:51.415008 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n7gp8" Dec 01 17:40:51 crc kubenswrapper[4868]: I1201 17:40:51.415378 4868 scope.go:117] "RemoveContainer" containerID="60d02cd775a696c94b7462cac0485d14f721e3f598750d90d4d357bd975de7ed" Dec 01 17:40:51 crc kubenswrapper[4868]: I1201 17:40:51.450537 4868 scope.go:117] "RemoveContainer" containerID="f682da9acdd7004061c5071ab0aa4e19f58db87989b657f262410acc96e9fce0" Dec 01 17:40:51 crc kubenswrapper[4868]: I1201 17:40:51.451591 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n7gp8"] Dec 01 17:40:51 crc kubenswrapper[4868]: I1201 17:40:51.457582 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n7gp8"] Dec 01 17:40:51 crc kubenswrapper[4868]: I1201 17:40:51.473564 4868 scope.go:117] "RemoveContainer" containerID="b33a0963b89e6f856f942806a8e3390f9242e947786b497279a1744dfe675f44" Dec 01 17:40:52 crc kubenswrapper[4868]: I1201 17:40:52.184221 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f0c4cd5-91cc-4683-9bf4-38e725958049" path="/var/lib/kubelet/pods/3f0c4cd5-91cc-4683-9bf4-38e725958049/volumes" Dec 01 17:40:52 crc kubenswrapper[4868]: I1201 17:40:52.435436 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rjfhx" podUID="e61d330e-2486-4ed9-84a5-6600251afb44" containerName="registry-server" containerID="cri-o://5d8680ee0f53548ea9c7df11ded7b08b91dc376772db5165e8b5df52dcf5cbc2" gracePeriod=2 Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.327891 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.435245 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sspb2\" (UniqueName: \"kubernetes.io/projected/e61d330e-2486-4ed9-84a5-6600251afb44-kube-api-access-sspb2\") pod \"e61d330e-2486-4ed9-84a5-6600251afb44\" (UID: \"e61d330e-2486-4ed9-84a5-6600251afb44\") " Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.435357 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e61d330e-2486-4ed9-84a5-6600251afb44-catalog-content\") pod \"e61d330e-2486-4ed9-84a5-6600251afb44\" (UID: \"e61d330e-2486-4ed9-84a5-6600251afb44\") " Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.435407 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e61d330e-2486-4ed9-84a5-6600251afb44-utilities\") pod \"e61d330e-2486-4ed9-84a5-6600251afb44\" (UID: \"e61d330e-2486-4ed9-84a5-6600251afb44\") " Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.436461 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e61d330e-2486-4ed9-84a5-6600251afb44-utilities" (OuterVolumeSpecName: "utilities") pod "e61d330e-2486-4ed9-84a5-6600251afb44" (UID: "e61d330e-2486-4ed9-84a5-6600251afb44"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.446891 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e61d330e-2486-4ed9-84a5-6600251afb44-kube-api-access-sspb2" (OuterVolumeSpecName: "kube-api-access-sspb2") pod "e61d330e-2486-4ed9-84a5-6600251afb44" (UID: "e61d330e-2486-4ed9-84a5-6600251afb44"). InnerVolumeSpecName "kube-api-access-sspb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.448926 4868 generic.go:334] "Generic (PLEG): container finished" podID="e61d330e-2486-4ed9-84a5-6600251afb44" containerID="5d8680ee0f53548ea9c7df11ded7b08b91dc376772db5165e8b5df52dcf5cbc2" exitCode=0 Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.448992 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjfhx" event={"ID":"e61d330e-2486-4ed9-84a5-6600251afb44","Type":"ContainerDied","Data":"5d8680ee0f53548ea9c7df11ded7b08b91dc376772db5165e8b5df52dcf5cbc2"} Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.449023 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjfhx" event={"ID":"e61d330e-2486-4ed9-84a5-6600251afb44","Type":"ContainerDied","Data":"53bbd2de1897a49cc884e038697634eae25eedd0006c8e3f9867cf166d9379c8"} Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.449042 4868 scope.go:117] "RemoveContainer" containerID="5d8680ee0f53548ea9c7df11ded7b08b91dc376772db5165e8b5df52dcf5cbc2" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.449197 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjfhx" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.488380 4868 scope.go:117] "RemoveContainer" containerID="90a4877ee7423aface54c4fa40a8874a7153c75fb20180ea42618faeac6a59c2" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.510520 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e61d330e-2486-4ed9-84a5-6600251afb44-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e61d330e-2486-4ed9-84a5-6600251afb44" (UID: "e61d330e-2486-4ed9-84a5-6600251afb44"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.512173 4868 scope.go:117] "RemoveContainer" containerID="0733de993a4eb0f5cf01faf6b52fd2d05c3a3a749bdcb1a5f3677da9e77cd786" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.530234 4868 scope.go:117] "RemoveContainer" containerID="5d8680ee0f53548ea9c7df11ded7b08b91dc376772db5165e8b5df52dcf5cbc2" Dec 01 17:40:53 crc kubenswrapper[4868]: E1201 17:40:53.530837 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d8680ee0f53548ea9c7df11ded7b08b91dc376772db5165e8b5df52dcf5cbc2\": container with ID starting with 5d8680ee0f53548ea9c7df11ded7b08b91dc376772db5165e8b5df52dcf5cbc2 not found: ID does not exist" containerID="5d8680ee0f53548ea9c7df11ded7b08b91dc376772db5165e8b5df52dcf5cbc2" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.530903 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d8680ee0f53548ea9c7df11ded7b08b91dc376772db5165e8b5df52dcf5cbc2"} err="failed to get container status \"5d8680ee0f53548ea9c7df11ded7b08b91dc376772db5165e8b5df52dcf5cbc2\": rpc error: code = NotFound desc = could not find container \"5d8680ee0f53548ea9c7df11ded7b08b91dc376772db5165e8b5df52dcf5cbc2\": container with ID starting with 5d8680ee0f53548ea9c7df11ded7b08b91dc376772db5165e8b5df52dcf5cbc2 not found: ID does not exist" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.530965 4868 scope.go:117] "RemoveContainer" containerID="90a4877ee7423aface54c4fa40a8874a7153c75fb20180ea42618faeac6a59c2" Dec 01 17:40:53 crc kubenswrapper[4868]: E1201 17:40:53.531490 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90a4877ee7423aface54c4fa40a8874a7153c75fb20180ea42618faeac6a59c2\": container with ID starting with 90a4877ee7423aface54c4fa40a8874a7153c75fb20180ea42618faeac6a59c2 not found: ID does not exist" containerID="90a4877ee7423aface54c4fa40a8874a7153c75fb20180ea42618faeac6a59c2" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.531581 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90a4877ee7423aface54c4fa40a8874a7153c75fb20180ea42618faeac6a59c2"} err="failed to get container status \"90a4877ee7423aface54c4fa40a8874a7153c75fb20180ea42618faeac6a59c2\": rpc error: code = NotFound desc = could not find container \"90a4877ee7423aface54c4fa40a8874a7153c75fb20180ea42618faeac6a59c2\": container with ID starting with 90a4877ee7423aface54c4fa40a8874a7153c75fb20180ea42618faeac6a59c2 not found: ID does not exist" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.531623 4868 scope.go:117] "RemoveContainer" containerID="0733de993a4eb0f5cf01faf6b52fd2d05c3a3a749bdcb1a5f3677da9e77cd786" Dec 01 17:40:53 crc kubenswrapper[4868]: E1201 17:40:53.532440 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0733de993a4eb0f5cf01faf6b52fd2d05c3a3a749bdcb1a5f3677da9e77cd786\": container with ID starting with 0733de993a4eb0f5cf01faf6b52fd2d05c3a3a749bdcb1a5f3677da9e77cd786 not found: ID does not exist" containerID="0733de993a4eb0f5cf01faf6b52fd2d05c3a3a749bdcb1a5f3677da9e77cd786" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.532469 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0733de993a4eb0f5cf01faf6b52fd2d05c3a3a749bdcb1a5f3677da9e77cd786"} err="failed to get container status \"0733de993a4eb0f5cf01faf6b52fd2d05c3a3a749bdcb1a5f3677da9e77cd786\": rpc error: code = NotFound desc = could not find container \"0733de993a4eb0f5cf01faf6b52fd2d05c3a3a749bdcb1a5f3677da9e77cd786\": container with ID starting with 0733de993a4eb0f5cf01faf6b52fd2d05c3a3a749bdcb1a5f3677da9e77cd786 not found: ID does not exist" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.537546 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sspb2\" (UniqueName: \"kubernetes.io/projected/e61d330e-2486-4ed9-84a5-6600251afb44-kube-api-access-sspb2\") on node \"crc\" DevicePath \"\"" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.537654 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e61d330e-2486-4ed9-84a5-6600251afb44-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.537674 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e61d330e-2486-4ed9-84a5-6600251afb44-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.779467 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rjfhx"] Dec 01 17:40:53 crc kubenswrapper[4868]: I1201 17:40:53.784801 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rjfhx"] Dec 01 17:40:54 crc kubenswrapper[4868]: I1201 17:40:54.189326 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e61d330e-2486-4ed9-84a5-6600251afb44" path="/var/lib/kubelet/pods/e61d330e-2486-4ed9-84a5-6600251afb44/volumes" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.492406 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rcxdx"] Dec 01 17:41:02 crc kubenswrapper[4868]: E1201 17:41:02.493255 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e61d330e-2486-4ed9-84a5-6600251afb44" containerName="registry-server" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.493267 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="e61d330e-2486-4ed9-84a5-6600251afb44" containerName="registry-server" Dec 01 17:41:02 crc kubenswrapper[4868]: E1201 17:41:02.493291 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e61d330e-2486-4ed9-84a5-6600251afb44" containerName="extract-content" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.493297 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="e61d330e-2486-4ed9-84a5-6600251afb44" containerName="extract-content" Dec 01 17:41:02 crc kubenswrapper[4868]: E1201 17:41:02.493327 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f0c4cd5-91cc-4683-9bf4-38e725958049" containerName="extract-content" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.493335 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f0c4cd5-91cc-4683-9bf4-38e725958049" containerName="extract-content" Dec 01 17:41:02 crc kubenswrapper[4868]: E1201 17:41:02.493348 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f0c4cd5-91cc-4683-9bf4-38e725958049" containerName="extract-utilities" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.493354 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f0c4cd5-91cc-4683-9bf4-38e725958049" containerName="extract-utilities" Dec 01 17:41:02 crc kubenswrapper[4868]: E1201 17:41:02.493369 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e61d330e-2486-4ed9-84a5-6600251afb44" containerName="extract-utilities" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.493376 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="e61d330e-2486-4ed9-84a5-6600251afb44" containerName="extract-utilities" Dec 01 17:41:02 crc kubenswrapper[4868]: E1201 17:41:02.493390 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f0c4cd5-91cc-4683-9bf4-38e725958049" containerName="registry-server" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.493397 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f0c4cd5-91cc-4683-9bf4-38e725958049" containerName="registry-server" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.493524 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="e61d330e-2486-4ed9-84a5-6600251afb44" containerName="registry-server" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.493543 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f0c4cd5-91cc-4683-9bf4-38e725958049" containerName="registry-server" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.494567 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rcxdx" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.520523 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rcxdx"] Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.586101 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a904bedc-9d01-4a00-82c1-6f2826a33d81-catalog-content\") pod \"redhat-operators-rcxdx\" (UID: \"a904bedc-9d01-4a00-82c1-6f2826a33d81\") " pod="openshift-marketplace/redhat-operators-rcxdx" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.586145 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a904bedc-9d01-4a00-82c1-6f2826a33d81-utilities\") pod \"redhat-operators-rcxdx\" (UID: \"a904bedc-9d01-4a00-82c1-6f2826a33d81\") " pod="openshift-marketplace/redhat-operators-rcxdx" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.586178 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j882h\" (UniqueName: \"kubernetes.io/projected/a904bedc-9d01-4a00-82c1-6f2826a33d81-kube-api-access-j882h\") pod \"redhat-operators-rcxdx\" (UID: \"a904bedc-9d01-4a00-82c1-6f2826a33d81\") " pod="openshift-marketplace/redhat-operators-rcxdx" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.687851 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a904bedc-9d01-4a00-82c1-6f2826a33d81-catalog-content\") pod \"redhat-operators-rcxdx\" (UID: \"a904bedc-9d01-4a00-82c1-6f2826a33d81\") " pod="openshift-marketplace/redhat-operators-rcxdx" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.688172 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a904bedc-9d01-4a00-82c1-6f2826a33d81-utilities\") pod \"redhat-operators-rcxdx\" (UID: \"a904bedc-9d01-4a00-82c1-6f2826a33d81\") " pod="openshift-marketplace/redhat-operators-rcxdx" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.688268 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j882h\" (UniqueName: \"kubernetes.io/projected/a904bedc-9d01-4a00-82c1-6f2826a33d81-kube-api-access-j882h\") pod \"redhat-operators-rcxdx\" (UID: \"a904bedc-9d01-4a00-82c1-6f2826a33d81\") " pod="openshift-marketplace/redhat-operators-rcxdx" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.688464 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a904bedc-9d01-4a00-82c1-6f2826a33d81-catalog-content\") pod \"redhat-operators-rcxdx\" (UID: \"a904bedc-9d01-4a00-82c1-6f2826a33d81\") " pod="openshift-marketplace/redhat-operators-rcxdx" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.688652 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a904bedc-9d01-4a00-82c1-6f2826a33d81-utilities\") pod \"redhat-operators-rcxdx\" (UID: \"a904bedc-9d01-4a00-82c1-6f2826a33d81\") " pod="openshift-marketplace/redhat-operators-rcxdx" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.721296 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j882h\" (UniqueName: \"kubernetes.io/projected/a904bedc-9d01-4a00-82c1-6f2826a33d81-kube-api-access-j882h\") pod \"redhat-operators-rcxdx\" (UID: \"a904bedc-9d01-4a00-82c1-6f2826a33d81\") " pod="openshift-marketplace/redhat-operators-rcxdx" Dec 01 17:41:02 crc kubenswrapper[4868]: I1201 17:41:02.815250 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rcxdx" Dec 01 17:41:03 crc kubenswrapper[4868]: I1201 17:41:03.282972 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rcxdx"] Dec 01 17:41:03 crc kubenswrapper[4868]: I1201 17:41:03.523269 4868 generic.go:334] "Generic (PLEG): container finished" podID="a904bedc-9d01-4a00-82c1-6f2826a33d81" containerID="0837f9c8eed3d6283798f1a4a6c489d02a657e9c90aec7de0982bcafafe2d648" exitCode=0 Dec 01 17:41:03 crc kubenswrapper[4868]: I1201 17:41:03.523343 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcxdx" event={"ID":"a904bedc-9d01-4a00-82c1-6f2826a33d81","Type":"ContainerDied","Data":"0837f9c8eed3d6283798f1a4a6c489d02a657e9c90aec7de0982bcafafe2d648"} Dec 01 17:41:03 crc kubenswrapper[4868]: I1201 17:41:03.523680 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcxdx" event={"ID":"a904bedc-9d01-4a00-82c1-6f2826a33d81","Type":"ContainerStarted","Data":"55ab0febac49e8b40d60f3af3602eecef4a3e8cefe8bd5df3e61e756d4c69875"} Dec 01 17:41:03 crc kubenswrapper[4868]: I1201 17:41:03.525322 4868 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 17:41:05 crc kubenswrapper[4868]: I1201 17:41:05.883564 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vjpd8"] Dec 01 17:41:05 crc kubenswrapper[4868]: I1201 17:41:05.890502 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-vjpd8" Dec 01 17:41:05 crc kubenswrapper[4868]: I1201 17:41:05.894175 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 01 17:41:05 crc kubenswrapper[4868]: I1201 17:41:05.900748 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 01 17:41:05 crc kubenswrapper[4868]: I1201 17:41:05.900969 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-lwbhh" Dec 01 17:41:05 crc kubenswrapper[4868]: I1201 17:41:05.903082 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 01 17:41:05 crc kubenswrapper[4868]: I1201 17:41:05.923007 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vjpd8"] Dec 01 17:41:05 crc kubenswrapper[4868]: I1201 17:41:05.941661 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qnlf\" (UniqueName: \"kubernetes.io/projected/7cafaee0-d17b-4e2e-a6c4-67382ce926c9-kube-api-access-4qnlf\") pod \"dnsmasq-dns-675f4bcbfc-vjpd8\" (UID: \"7cafaee0-d17b-4e2e-a6c4-67382ce926c9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vjpd8" Dec 01 17:41:05 crc kubenswrapper[4868]: I1201 17:41:05.941727 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cafaee0-d17b-4e2e-a6c4-67382ce926c9-config\") pod \"dnsmasq-dns-675f4bcbfc-vjpd8\" (UID: \"7cafaee0-d17b-4e2e-a6c4-67382ce926c9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vjpd8" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.018843 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6r4nr"] Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.020066 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.023792 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.068864 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cec39435-608a-497c-8802-0cc4639599bf-config\") pod \"dnsmasq-dns-78dd6ddcc-6r4nr\" (UID: \"cec39435-608a-497c-8802-0cc4639599bf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.068965 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cafaee0-d17b-4e2e-a6c4-67382ce926c9-config\") pod \"dnsmasq-dns-675f4bcbfc-vjpd8\" (UID: \"7cafaee0-d17b-4e2e-a6c4-67382ce926c9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vjpd8" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.069029 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cec39435-608a-497c-8802-0cc4639599bf-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-6r4nr\" (UID: \"cec39435-608a-497c-8802-0cc4639599bf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.069279 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j7t7\" (UniqueName: \"kubernetes.io/projected/cec39435-608a-497c-8802-0cc4639599bf-kube-api-access-6j7t7\") pod \"dnsmasq-dns-78dd6ddcc-6r4nr\" (UID: \"cec39435-608a-497c-8802-0cc4639599bf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.069321 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qnlf\" (UniqueName: \"kubernetes.io/projected/7cafaee0-d17b-4e2e-a6c4-67382ce926c9-kube-api-access-4qnlf\") pod \"dnsmasq-dns-675f4bcbfc-vjpd8\" (UID: \"7cafaee0-d17b-4e2e-a6c4-67382ce926c9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vjpd8" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.072591 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cafaee0-d17b-4e2e-a6c4-67382ce926c9-config\") pod \"dnsmasq-dns-675f4bcbfc-vjpd8\" (UID: \"7cafaee0-d17b-4e2e-a6c4-67382ce926c9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vjpd8" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.094490 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6r4nr"] Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.108653 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qnlf\" (UniqueName: \"kubernetes.io/projected/7cafaee0-d17b-4e2e-a6c4-67382ce926c9-kube-api-access-4qnlf\") pod \"dnsmasq-dns-675f4bcbfc-vjpd8\" (UID: \"7cafaee0-d17b-4e2e-a6c4-67382ce926c9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vjpd8" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.171131 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j7t7\" (UniqueName: \"kubernetes.io/projected/cec39435-608a-497c-8802-0cc4639599bf-kube-api-access-6j7t7\") pod \"dnsmasq-dns-78dd6ddcc-6r4nr\" (UID: \"cec39435-608a-497c-8802-0cc4639599bf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.171187 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cec39435-608a-497c-8802-0cc4639599bf-config\") pod \"dnsmasq-dns-78dd6ddcc-6r4nr\" (UID: \"cec39435-608a-497c-8802-0cc4639599bf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.171213 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cec39435-608a-497c-8802-0cc4639599bf-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-6r4nr\" (UID: \"cec39435-608a-497c-8802-0cc4639599bf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.172198 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cec39435-608a-497c-8802-0cc4639599bf-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-6r4nr\" (UID: \"cec39435-608a-497c-8802-0cc4639599bf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.172343 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cec39435-608a-497c-8802-0cc4639599bf-config\") pod \"dnsmasq-dns-78dd6ddcc-6r4nr\" (UID: \"cec39435-608a-497c-8802-0cc4639599bf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.189865 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j7t7\" (UniqueName: \"kubernetes.io/projected/cec39435-608a-497c-8802-0cc4639599bf-kube-api-access-6j7t7\") pod \"dnsmasq-dns-78dd6ddcc-6r4nr\" (UID: \"cec39435-608a-497c-8802-0cc4639599bf\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.223515 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-vjpd8" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.353008 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.665075 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vjpd8"] Dec 01 17:41:06 crc kubenswrapper[4868]: W1201 17:41:06.675421 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7cafaee0_d17b_4e2e_a6c4_67382ce926c9.slice/crio-94248d23982b57cb9b991a48f4a72075dd001a8d497f5fda8f00670f1fa09d83 WatchSource:0}: Error finding container 94248d23982b57cb9b991a48f4a72075dd001a8d497f5fda8f00670f1fa09d83: Status 404 returned error can't find the container with id 94248d23982b57cb9b991a48f4a72075dd001a8d497f5fda8f00670f1fa09d83 Dec 01 17:41:06 crc kubenswrapper[4868]: I1201 17:41:06.814275 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6r4nr"] Dec 01 17:41:07 crc kubenswrapper[4868]: I1201 17:41:07.561649 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-vjpd8" event={"ID":"7cafaee0-d17b-4e2e-a6c4-67382ce926c9","Type":"ContainerStarted","Data":"94248d23982b57cb9b991a48f4a72075dd001a8d497f5fda8f00670f1fa09d83"} Dec 01 17:41:07 crc kubenswrapper[4868]: I1201 17:41:07.562688 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" event={"ID":"cec39435-608a-497c-8802-0cc4639599bf","Type":"ContainerStarted","Data":"b1447c6951c2d423dc8c7a8086cccd92c9eba647690079d971d758e6867f71c5"} Dec 01 17:41:08 crc kubenswrapper[4868]: I1201 17:41:08.985727 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vjpd8"] Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.043329 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7ch5d"] Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.045007 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.080454 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7ch5d"] Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.129178 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ftkk\" (UniqueName: \"kubernetes.io/projected/6d71bfbf-8024-47e4-860c-c1cb363bb185-kube-api-access-6ftkk\") pod \"dnsmasq-dns-666b6646f7-7ch5d\" (UID: \"6d71bfbf-8024-47e4-860c-c1cb363bb185\") " pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.129233 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d71bfbf-8024-47e4-860c-c1cb363bb185-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7ch5d\" (UID: \"6d71bfbf-8024-47e4-860c-c1cb363bb185\") " pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.129262 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d71bfbf-8024-47e4-860c-c1cb363bb185-config\") pod \"dnsmasq-dns-666b6646f7-7ch5d\" (UID: \"6d71bfbf-8024-47e4-860c-c1cb363bb185\") " pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.233735 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ftkk\" (UniqueName: \"kubernetes.io/projected/6d71bfbf-8024-47e4-860c-c1cb363bb185-kube-api-access-6ftkk\") pod \"dnsmasq-dns-666b6646f7-7ch5d\" (UID: \"6d71bfbf-8024-47e4-860c-c1cb363bb185\") " pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.237520 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d71bfbf-8024-47e4-860c-c1cb363bb185-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7ch5d\" (UID: \"6d71bfbf-8024-47e4-860c-c1cb363bb185\") " pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.237591 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d71bfbf-8024-47e4-860c-c1cb363bb185-config\") pod \"dnsmasq-dns-666b6646f7-7ch5d\" (UID: \"6d71bfbf-8024-47e4-860c-c1cb363bb185\") " pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.239102 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d71bfbf-8024-47e4-860c-c1cb363bb185-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7ch5d\" (UID: \"6d71bfbf-8024-47e4-860c-c1cb363bb185\") " pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.239237 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d71bfbf-8024-47e4-860c-c1cb363bb185-config\") pod \"dnsmasq-dns-666b6646f7-7ch5d\" (UID: \"6d71bfbf-8024-47e4-860c-c1cb363bb185\") " pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.278565 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ftkk\" (UniqueName: \"kubernetes.io/projected/6d71bfbf-8024-47e4-860c-c1cb363bb185-kube-api-access-6ftkk\") pod \"dnsmasq-dns-666b6646f7-7ch5d\" (UID: \"6d71bfbf-8024-47e4-860c-c1cb363bb185\") " pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.366117 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.419213 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6r4nr"] Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.463595 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lr5ft"] Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.464897 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.490898 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lr5ft"] Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.544860 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba10e6ec-3b28-462a-b625-a79d63eb5f92-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-lr5ft\" (UID: \"ba10e6ec-3b28-462a-b625-a79d63eb5f92\") " pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.544922 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba10e6ec-3b28-462a-b625-a79d63eb5f92-config\") pod \"dnsmasq-dns-57d769cc4f-lr5ft\" (UID: \"ba10e6ec-3b28-462a-b625-a79d63eb5f92\") " pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.544964 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ghsb\" (UniqueName: \"kubernetes.io/projected/ba10e6ec-3b28-462a-b625-a79d63eb5f92-kube-api-access-8ghsb\") pod \"dnsmasq-dns-57d769cc4f-lr5ft\" (UID: \"ba10e6ec-3b28-462a-b625-a79d63eb5f92\") " pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.646494 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba10e6ec-3b28-462a-b625-a79d63eb5f92-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-lr5ft\" (UID: \"ba10e6ec-3b28-462a-b625-a79d63eb5f92\") " pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.646552 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba10e6ec-3b28-462a-b625-a79d63eb5f92-config\") pod \"dnsmasq-dns-57d769cc4f-lr5ft\" (UID: \"ba10e6ec-3b28-462a-b625-a79d63eb5f92\") " pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.646580 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ghsb\" (UniqueName: \"kubernetes.io/projected/ba10e6ec-3b28-462a-b625-a79d63eb5f92-kube-api-access-8ghsb\") pod \"dnsmasq-dns-57d769cc4f-lr5ft\" (UID: \"ba10e6ec-3b28-462a-b625-a79d63eb5f92\") " pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.648205 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba10e6ec-3b28-462a-b625-a79d63eb5f92-config\") pod \"dnsmasq-dns-57d769cc4f-lr5ft\" (UID: \"ba10e6ec-3b28-462a-b625-a79d63eb5f92\") " pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.649402 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba10e6ec-3b28-462a-b625-a79d63eb5f92-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-lr5ft\" (UID: \"ba10e6ec-3b28-462a-b625-a79d63eb5f92\") " pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.667651 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ghsb\" (UniqueName: \"kubernetes.io/projected/ba10e6ec-3b28-462a-b625-a79d63eb5f92-kube-api-access-8ghsb\") pod \"dnsmasq-dns-57d769cc4f-lr5ft\" (UID: \"ba10e6ec-3b28-462a-b625-a79d63eb5f92\") " pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" Dec 01 17:41:09 crc kubenswrapper[4868]: I1201 17:41:09.799870 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.226930 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.229414 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.236258 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.236429 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.236956 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-94wgt" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.237123 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.237248 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.238056 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.238174 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.264845 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.275805 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7ch5d"] Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.333648 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lr5ft"] Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.357606 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.357647 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.357670 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.357696 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-server-conf\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.357813 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.358071 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.358194 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/70aee7bf-d819-4355-a98c-4a106f3e6f16-pod-info\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.358249 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/70aee7bf-d819-4355-a98c-4a106f3e6f16-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.358498 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgls7\" (UniqueName: \"kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-kube-api-access-pgls7\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.358581 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-config-data\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.358616 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.460214 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgls7\" (UniqueName: \"kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-kube-api-access-pgls7\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.460324 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-config-data\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.460373 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.460444 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.460461 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.460480 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.461068 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.461245 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.461326 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-server-conf\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.461405 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.461444 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.461480 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/70aee7bf-d819-4355-a98c-4a106f3e6f16-pod-info\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.461515 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/70aee7bf-d819-4355-a98c-4a106f3e6f16-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.461707 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.463487 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-config-data\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.463507 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.463772 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-server-conf\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.466712 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/70aee7bf-d819-4355-a98c-4a106f3e6f16-pod-info\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.468461 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.469084 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/70aee7bf-d819-4355-a98c-4a106f3e6f16-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.475776 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.479674 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgls7\" (UniqueName: \"kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-kube-api-access-pgls7\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.485603 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.573396 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.574958 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.578278 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.626327 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.626700 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.626776 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.626706 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.627218 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.629901 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.631926 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.632025 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-v799m" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.665230 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3ff4f488-06c8-4ceb-848a-4ba68b334752-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.665266 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.665311 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.665348 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgtg8\" (UniqueName: \"kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-kube-api-access-tgtg8\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.665367 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.665394 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.665420 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3ff4f488-06c8-4ceb-848a-4ba68b334752-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.665447 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.665466 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.665485 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.665514 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.767448 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.767502 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3ff4f488-06c8-4ceb-848a-4ba68b334752-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.767520 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.767560 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.767591 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgtg8\" (UniqueName: \"kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-kube-api-access-tgtg8\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.767611 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.767638 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.767668 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3ff4f488-06c8-4ceb-848a-4ba68b334752-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.767695 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.767716 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.767732 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.768980 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.769692 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.769979 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.770071 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.770442 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.770458 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.773004 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3ff4f488-06c8-4ceb-848a-4ba68b334752-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.793290 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgtg8\" (UniqueName: \"kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-kube-api-access-tgtg8\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.793873 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3ff4f488-06c8-4ceb-848a-4ba68b334752-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.794016 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.800418 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.818554 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:10 crc kubenswrapper[4868]: I1201 17:41:10.976050 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.045811 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.047859 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.061808 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.066057 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-s5tv6" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.066239 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.070767 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.071321 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.072792 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.207830 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6h6t\" (UniqueName: \"kubernetes.io/projected/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-kube-api-access-h6h6t\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.207892 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.207987 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.208013 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-config-data-default\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.208030 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.208059 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.208074 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-kolla-config\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.208095 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.309856 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.309921 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-config-data-default\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.309963 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.310048 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.310069 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-kolla-config\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.310095 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.310160 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6h6t\" (UniqueName: \"kubernetes.io/projected/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-kube-api-access-h6h6t\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.310181 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.310626 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.310810 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.311042 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-config-data-default\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.311132 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-kolla-config\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.311996 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.316553 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.320412 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.332659 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6h6t\" (UniqueName: \"kubernetes.io/projected/68f3c214-92ad-43c0-b06a-33c6f1f90cfb-kube-api-access-h6h6t\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.352430 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"68f3c214-92ad-43c0-b06a-33c6f1f90cfb\") " pod="openstack/openstack-galera-0" Dec 01 17:41:12 crc kubenswrapper[4868]: I1201 17:41:12.423686 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.289929 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.291441 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.293564 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.294248 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.296087 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.312110 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.322731 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.323978 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.331555 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.358766 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.358893 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.361259 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-l4mbx" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.361828 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-lhc2f" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.428422 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cc48650c-5f3d-4c0a-b570-c7eb8615d504-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.428739 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e-kolla-config\") pod \"memcached-0\" (UID: \"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e\") " pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.428767 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cc48650c-5f3d-4c0a-b570-c7eb8615d504-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.428805 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc48650c-5f3d-4c0a-b570-c7eb8615d504-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.428828 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc48650c-5f3d-4c0a-b570-c7eb8615d504-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.428849 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cc48650c-5f3d-4c0a-b570-c7eb8615d504-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.428867 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbtxl\" (UniqueName: \"kubernetes.io/projected/4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e-kube-api-access-rbtxl\") pod \"memcached-0\" (UID: \"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e\") " pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.429052 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc48650c-5f3d-4c0a-b570-c7eb8615d504-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.429160 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e\") " pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.429206 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e-config-data\") pod \"memcached-0\" (UID: \"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e\") " pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.429236 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.429308 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e\") " pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.429375 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6q9g\" (UniqueName: \"kubernetes.io/projected/cc48650c-5f3d-4c0a-b570-c7eb8615d504-kube-api-access-p6q9g\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.530591 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e-kolla-config\") pod \"memcached-0\" (UID: \"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e\") " pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.530723 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cc48650c-5f3d-4c0a-b570-c7eb8615d504-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.530781 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc48650c-5f3d-4c0a-b570-c7eb8615d504-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.530814 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc48650c-5f3d-4c0a-b570-c7eb8615d504-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.530842 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cc48650c-5f3d-4c0a-b570-c7eb8615d504-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.530866 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbtxl\" (UniqueName: \"kubernetes.io/projected/4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e-kube-api-access-rbtxl\") pod \"memcached-0\" (UID: \"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e\") " pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.530954 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc48650c-5f3d-4c0a-b570-c7eb8615d504-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.530995 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e\") " pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.531022 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e-config-data\") pod \"memcached-0\" (UID: \"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e\") " pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.531054 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.531090 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e\") " pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.531134 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6q9g\" (UniqueName: \"kubernetes.io/projected/cc48650c-5f3d-4c0a-b570-c7eb8615d504-kube-api-access-p6q9g\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.531167 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cc48650c-5f3d-4c0a-b570-c7eb8615d504-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.531518 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.531675 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e-kolla-config\") pod \"memcached-0\" (UID: \"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e\") " pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.531964 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cc48650c-5f3d-4c0a-b570-c7eb8615d504-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.532081 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cc48650c-5f3d-4c0a-b570-c7eb8615d504-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.532885 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e-config-data\") pod \"memcached-0\" (UID: \"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e\") " pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.533142 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cc48650c-5f3d-4c0a-b570-c7eb8615d504-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.533401 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc48650c-5f3d-4c0a-b570-c7eb8615d504-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.547204 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc48650c-5f3d-4c0a-b570-c7eb8615d504-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.551354 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6q9g\" (UniqueName: \"kubernetes.io/projected/cc48650c-5f3d-4c0a-b570-c7eb8615d504-kube-api-access-p6q9g\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.551627 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbtxl\" (UniqueName: \"kubernetes.io/projected/4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e-kube-api-access-rbtxl\") pod \"memcached-0\" (UID: \"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e\") " pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.553578 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e\") " pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.554151 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc48650c-5f3d-4c0a-b570-c7eb8615d504-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.571190 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e\") " pod="openstack/memcached-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.574904 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"cc48650c-5f3d-4c0a-b570-c7eb8615d504\") " pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.662678 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 17:41:13 crc kubenswrapper[4868]: I1201 17:41:13.678756 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 17:41:15 crc kubenswrapper[4868]: I1201 17:41:15.453492 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 17:41:15 crc kubenswrapper[4868]: I1201 17:41:15.454925 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 17:41:15 crc kubenswrapper[4868]: I1201 17:41:15.458510 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-9pgtm" Dec 01 17:41:15 crc kubenswrapper[4868]: I1201 17:41:15.516754 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 17:41:15 crc kubenswrapper[4868]: I1201 17:41:15.565266 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kktwt\" (UniqueName: \"kubernetes.io/projected/d5daf33c-c3fa-4f86-a422-93e7afa1afaa-kube-api-access-kktwt\") pod \"kube-state-metrics-0\" (UID: \"d5daf33c-c3fa-4f86-a422-93e7afa1afaa\") " pod="openstack/kube-state-metrics-0" Dec 01 17:41:15 crc kubenswrapper[4868]: I1201 17:41:15.667147 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kktwt\" (UniqueName: \"kubernetes.io/projected/d5daf33c-c3fa-4f86-a422-93e7afa1afaa-kube-api-access-kktwt\") pod \"kube-state-metrics-0\" (UID: \"d5daf33c-c3fa-4f86-a422-93e7afa1afaa\") " pod="openstack/kube-state-metrics-0" Dec 01 17:41:15 crc kubenswrapper[4868]: I1201 17:41:15.687401 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kktwt\" (UniqueName: \"kubernetes.io/projected/d5daf33c-c3fa-4f86-a422-93e7afa1afaa-kube-api-access-kktwt\") pod \"kube-state-metrics-0\" (UID: \"d5daf33c-c3fa-4f86-a422-93e7afa1afaa\") " pod="openstack/kube-state-metrics-0" Dec 01 17:41:15 crc kubenswrapper[4868]: I1201 17:41:15.774246 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.638520 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-z4f4z"] Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.641072 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.648975 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.649141 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-mmd77" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.652640 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.664175 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z4f4z"] Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.710283 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-rzgnl"] Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.715754 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.717272 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65d7a832-930e-4103-90f6-dbc5de8c1ece-combined-ca-bundle\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.717658 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftsgx\" (UniqueName: \"kubernetes.io/projected/65d7a832-930e-4103-90f6-dbc5de8c1ece-kube-api-access-ftsgx\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.717793 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/65d7a832-930e-4103-90f6-dbc5de8c1ece-var-run-ovn\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.718137 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/65d7a832-930e-4103-90f6-dbc5de8c1ece-var-run\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.718231 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/65d7a832-930e-4103-90f6-dbc5de8c1ece-ovn-controller-tls-certs\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.718337 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65d7a832-930e-4103-90f6-dbc5de8c1ece-scripts\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.718406 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/65d7a832-930e-4103-90f6-dbc5de8c1ece-var-log-ovn\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.730445 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rzgnl"] Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.819632 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/8cca4bfb-cddc-425e-abe2-6debea4aa92a-etc-ovs\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.819706 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwkn4\" (UniqueName: \"kubernetes.io/projected/8cca4bfb-cddc-425e-abe2-6debea4aa92a-kube-api-access-pwkn4\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.819749 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/65d7a832-930e-4103-90f6-dbc5de8c1ece-var-run\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.819778 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8cca4bfb-cddc-425e-abe2-6debea4aa92a-var-run\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.819809 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/65d7a832-930e-4103-90f6-dbc5de8c1ece-ovn-controller-tls-certs\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.819837 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/8cca4bfb-cddc-425e-abe2-6debea4aa92a-var-lib\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.819865 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8cca4bfb-cddc-425e-abe2-6debea4aa92a-scripts\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.819903 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65d7a832-930e-4103-90f6-dbc5de8c1ece-scripts\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.819971 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/65d7a832-930e-4103-90f6-dbc5de8c1ece-var-log-ovn\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.819999 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65d7a832-930e-4103-90f6-dbc5de8c1ece-combined-ca-bundle\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.820024 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftsgx\" (UniqueName: \"kubernetes.io/projected/65d7a832-930e-4103-90f6-dbc5de8c1ece-kube-api-access-ftsgx\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.820052 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/65d7a832-930e-4103-90f6-dbc5de8c1ece-var-run-ovn\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.820074 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/8cca4bfb-cddc-425e-abe2-6debea4aa92a-var-log\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.820457 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/65d7a832-930e-4103-90f6-dbc5de8c1ece-var-run\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.820598 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/65d7a832-930e-4103-90f6-dbc5de8c1ece-var-log-ovn\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.821021 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/65d7a832-930e-4103-90f6-dbc5de8c1ece-var-run-ovn\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.822909 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65d7a832-930e-4103-90f6-dbc5de8c1ece-scripts\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.830926 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/65d7a832-930e-4103-90f6-dbc5de8c1ece-ovn-controller-tls-certs\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.837465 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65d7a832-930e-4103-90f6-dbc5de8c1ece-combined-ca-bundle\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.837550 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftsgx\" (UniqueName: \"kubernetes.io/projected/65d7a832-930e-4103-90f6-dbc5de8c1ece-kube-api-access-ftsgx\") pod \"ovn-controller-z4f4z\" (UID: \"65d7a832-930e-4103-90f6-dbc5de8c1ece\") " pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.922167 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/8cca4bfb-cddc-425e-abe2-6debea4aa92a-etc-ovs\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.922909 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwkn4\" (UniqueName: \"kubernetes.io/projected/8cca4bfb-cddc-425e-abe2-6debea4aa92a-kube-api-access-pwkn4\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.923101 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8cca4bfb-cddc-425e-abe2-6debea4aa92a-var-run\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.923237 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/8cca4bfb-cddc-425e-abe2-6debea4aa92a-var-lib\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.923368 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8cca4bfb-cddc-425e-abe2-6debea4aa92a-scripts\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.923467 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/8cca4bfb-cddc-425e-abe2-6debea4aa92a-var-lib\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.923247 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8cca4bfb-cddc-425e-abe2-6debea4aa92a-var-run\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.922503 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/8cca4bfb-cddc-425e-abe2-6debea4aa92a-etc-ovs\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.925719 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8cca4bfb-cddc-425e-abe2-6debea4aa92a-scripts\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.926008 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/8cca4bfb-cddc-425e-abe2-6debea4aa92a-var-log\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.926326 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/8cca4bfb-cddc-425e-abe2-6debea4aa92a-var-log\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.955186 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwkn4\" (UniqueName: \"kubernetes.io/projected/8cca4bfb-cddc-425e-abe2-6debea4aa92a-kube-api-access-pwkn4\") pod \"ovn-controller-ovs-rzgnl\" (UID: \"8cca4bfb-cddc-425e-abe2-6debea4aa92a\") " pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:18 crc kubenswrapper[4868]: I1201 17:41:18.960079 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.039219 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.508098 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.509655 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.511739 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.512172 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.515687 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.515698 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-h8dfh" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.515876 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.520675 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.639108 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.639176 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.639368 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-config\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.639478 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.639509 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.639589 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4xgd\" (UniqueName: \"kubernetes.io/projected/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-kube-api-access-b4xgd\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.639701 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.639880 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.742020 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4xgd\" (UniqueName: \"kubernetes.io/projected/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-kube-api-access-b4xgd\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.742079 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.742126 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.742163 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.742223 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.742268 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-config\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.742300 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.742322 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.742562 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.742967 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.743361 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-config\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.743857 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.747267 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.747513 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.748061 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.758773 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4xgd\" (UniqueName: \"kubernetes.io/projected/89b376d8-f72b-4c03-a2f1-3fcf6ff20336-kube-api-access-b4xgd\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.765480 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"89b376d8-f72b-4c03-a2f1-3fcf6ff20336\") " pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.770100 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" event={"ID":"6d71bfbf-8024-47e4-860c-c1cb363bb185","Type":"ContainerStarted","Data":"33dfa299d9ddfdb9a56c6f14a13020d5cb737c8b566c54f5bb6337ff2b0b4230"} Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.771735 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" event={"ID":"ba10e6ec-3b28-462a-b625-a79d63eb5f92","Type":"ContainerStarted","Data":"e1f514e1b0ed9e46125087bafbf57c9a6ae9b0426bcdbb12242cb45cadc54c35"} Dec 01 17:41:19 crc kubenswrapper[4868]: I1201 17:41:19.833483 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 17:41:20 crc kubenswrapper[4868]: I1201 17:41:20.746788 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.067581 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.069748 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.074042 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-hvbjv" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.074287 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.074562 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.081746 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.085552 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.193314 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faa33685-51f4-46ce-bdc7-ef9a2467f482-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.193704 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faa33685-51f4-46ce-bdc7-ef9a2467f482-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.193737 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75sk2\" (UniqueName: \"kubernetes.io/projected/faa33685-51f4-46ce-bdc7-ef9a2467f482-kube-api-access-75sk2\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.193796 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/faa33685-51f4-46ce-bdc7-ef9a2467f482-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.193841 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/faa33685-51f4-46ce-bdc7-ef9a2467f482-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.193868 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faa33685-51f4-46ce-bdc7-ef9a2467f482-config\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.193923 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/faa33685-51f4-46ce-bdc7-ef9a2467f482-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.193982 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.296140 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/faa33685-51f4-46ce-bdc7-ef9a2467f482-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.296235 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/faa33685-51f4-46ce-bdc7-ef9a2467f482-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.296258 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faa33685-51f4-46ce-bdc7-ef9a2467f482-config\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.296312 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/faa33685-51f4-46ce-bdc7-ef9a2467f482-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.296351 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.296373 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faa33685-51f4-46ce-bdc7-ef9a2467f482-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.296407 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faa33685-51f4-46ce-bdc7-ef9a2467f482-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.296434 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75sk2\" (UniqueName: \"kubernetes.io/projected/faa33685-51f4-46ce-bdc7-ef9a2467f482-kube-api-access-75sk2\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.297908 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/faa33685-51f4-46ce-bdc7-ef9a2467f482-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.299103 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faa33685-51f4-46ce-bdc7-ef9a2467f482-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.299811 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faa33685-51f4-46ce-bdc7-ef9a2467f482-config\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.300124 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.306409 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/faa33685-51f4-46ce-bdc7-ef9a2467f482-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.309825 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faa33685-51f4-46ce-bdc7-ef9a2467f482-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.313852 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/faa33685-51f4-46ce-bdc7-ef9a2467f482-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.316156 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75sk2\" (UniqueName: \"kubernetes.io/projected/faa33685-51f4-46ce-bdc7-ef9a2467f482-kube-api-access-75sk2\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.323085 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"faa33685-51f4-46ce-bdc7-ef9a2467f482\") " pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:22 crc kubenswrapper[4868]: I1201 17:41:22.403701 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 17:41:24 crc kubenswrapper[4868]: I1201 17:41:24.842587 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"70aee7bf-d819-4355-a98c-4a106f3e6f16","Type":"ContainerStarted","Data":"36fd322c034c02e8064e5407dbf69bd923bb67d3ba40e2c1eadad630a25cbed2"} Dec 01 17:41:31 crc kubenswrapper[4868]: I1201 17:41:31.511977 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q7782"] Dec 01 17:41:31 crc kubenswrapper[4868]: I1201 17:41:31.514750 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:41:31 crc kubenswrapper[4868]: I1201 17:41:31.531425 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q7782"] Dec 01 17:41:31 crc kubenswrapper[4868]: I1201 17:41:31.590610 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-utilities\") pod \"certified-operators-q7782\" (UID: \"d7e2ea14-2a0a-418c-ab25-4a588d728f5a\") " pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:41:31 crc kubenswrapper[4868]: I1201 17:41:31.590686 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-catalog-content\") pod \"certified-operators-q7782\" (UID: \"d7e2ea14-2a0a-418c-ab25-4a588d728f5a\") " pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:41:31 crc kubenswrapper[4868]: I1201 17:41:31.590772 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktplj\" (UniqueName: \"kubernetes.io/projected/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-kube-api-access-ktplj\") pod \"certified-operators-q7782\" (UID: \"d7e2ea14-2a0a-418c-ab25-4a588d728f5a\") " pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:41:31 crc kubenswrapper[4868]: I1201 17:41:31.692779 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktplj\" (UniqueName: \"kubernetes.io/projected/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-kube-api-access-ktplj\") pod \"certified-operators-q7782\" (UID: \"d7e2ea14-2a0a-418c-ab25-4a588d728f5a\") " pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:41:31 crc kubenswrapper[4868]: I1201 17:41:31.692860 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-utilities\") pod \"certified-operators-q7782\" (UID: \"d7e2ea14-2a0a-418c-ab25-4a588d728f5a\") " pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:41:31 crc kubenswrapper[4868]: I1201 17:41:31.692911 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-catalog-content\") pod \"certified-operators-q7782\" (UID: \"d7e2ea14-2a0a-418c-ab25-4a588d728f5a\") " pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:41:31 crc kubenswrapper[4868]: I1201 17:41:31.693539 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-catalog-content\") pod \"certified-operators-q7782\" (UID: \"d7e2ea14-2a0a-418c-ab25-4a588d728f5a\") " pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:41:31 crc kubenswrapper[4868]: I1201 17:41:31.693637 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-utilities\") pod \"certified-operators-q7782\" (UID: \"d7e2ea14-2a0a-418c-ab25-4a588d728f5a\") " pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:41:31 crc kubenswrapper[4868]: I1201 17:41:31.714484 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktplj\" (UniqueName: \"kubernetes.io/projected/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-kube-api-access-ktplj\") pod \"certified-operators-q7782\" (UID: \"d7e2ea14-2a0a-418c-ab25-4a588d728f5a\") " pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:41:31 crc kubenswrapper[4868]: I1201 17:41:31.832224 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:41:35 crc kubenswrapper[4868]: E1201 17:41:35.929499 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 17:41:35 crc kubenswrapper[4868]: E1201 17:41:35.930439 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6j7t7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-6r4nr_openstack(cec39435-608a-497c-8802-0cc4639599bf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 17:41:35 crc kubenswrapper[4868]: E1201 17:41:35.931603 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" podUID="cec39435-608a-497c-8802-0cc4639599bf" Dec 01 17:41:36 crc kubenswrapper[4868]: E1201 17:41:36.007861 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 17:41:36 crc kubenswrapper[4868]: E1201 17:41:36.008118 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4qnlf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-vjpd8_openstack(7cafaee0-d17b-4e2e-a6c4-67382ce926c9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 17:41:36 crc kubenswrapper[4868]: E1201 17:41:36.009570 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-vjpd8" podUID="7cafaee0-d17b-4e2e-a6c4-67382ce926c9" Dec 01 17:41:36 crc kubenswrapper[4868]: E1201 17:41:36.014934 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 17:41:36 crc kubenswrapper[4868]: E1201 17:41:36.015830 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6ftkk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-7ch5d_openstack(6d71bfbf-8024-47e4-860c-c1cb363bb185): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 17:41:36 crc kubenswrapper[4868]: E1201 17:41:36.017173 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" podUID="6d71bfbf-8024-47e4-860c-c1cb363bb185" Dec 01 17:41:36 crc kubenswrapper[4868]: I1201 17:41:36.537301 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 17:41:36 crc kubenswrapper[4868]: W1201 17:41:36.568552 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d1ee35e_4e53_46a0_a1d9_86eeb1f9c15e.slice/crio-547b2ef157802e655d361422a401044687031935fdf89252ba6405f0f7ebd7f4 WatchSource:0}: Error finding container 547b2ef157802e655d361422a401044687031935fdf89252ba6405f0f7ebd7f4: Status 404 returned error can't find the container with id 547b2ef157802e655d361422a401044687031935fdf89252ba6405f0f7ebd7f4 Dec 01 17:41:36 crc kubenswrapper[4868]: I1201 17:41:36.573348 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 17:41:36 crc kubenswrapper[4868]: W1201 17:41:36.579361 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68f3c214_92ad_43c0_b06a_33c6f1f90cfb.slice/crio-ee85449f82acdda2d35787c855bebc900d68ca1fa6d819f9b2b2cf2905782477 WatchSource:0}: Error finding container ee85449f82acdda2d35787c855bebc900d68ca1fa6d819f9b2b2cf2905782477: Status 404 returned error can't find the container with id ee85449f82acdda2d35787c855bebc900d68ca1fa6d819f9b2b2cf2905782477 Dec 01 17:41:36 crc kubenswrapper[4868]: I1201 17:41:36.600524 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 17:41:36 crc kubenswrapper[4868]: I1201 17:41:36.930415 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3ff4f488-06c8-4ceb-848a-4ba68b334752","Type":"ContainerStarted","Data":"e98cc91c9e8fb7ee706b45be8b9b3fbbbfe6e4dcd75eb3638b68007528d4acda"} Dec 01 17:41:36 crc kubenswrapper[4868]: I1201 17:41:36.931794 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e","Type":"ContainerStarted","Data":"547b2ef157802e655d361422a401044687031935fdf89252ba6405f0f7ebd7f4"} Dec 01 17:41:36 crc kubenswrapper[4868]: I1201 17:41:36.936108 4868 generic.go:334] "Generic (PLEG): container finished" podID="a904bedc-9d01-4a00-82c1-6f2826a33d81" containerID="05a6d9b87f176f7ce3ef72d22dd29d62ca8f163ceb949a6c7565c9f8d8892cc4" exitCode=0 Dec 01 17:41:36 crc kubenswrapper[4868]: I1201 17:41:36.936180 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcxdx" event={"ID":"a904bedc-9d01-4a00-82c1-6f2826a33d81","Type":"ContainerDied","Data":"05a6d9b87f176f7ce3ef72d22dd29d62ca8f163ceb949a6c7565c9f8d8892cc4"} Dec 01 17:41:36 crc kubenswrapper[4868]: I1201 17:41:36.937782 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"68f3c214-92ad-43c0-b06a-33c6f1f90cfb","Type":"ContainerStarted","Data":"ee85449f82acdda2d35787c855bebc900d68ca1fa6d819f9b2b2cf2905782477"} Dec 01 17:41:36 crc kubenswrapper[4868]: I1201 17:41:36.940752 4868 generic.go:334] "Generic (PLEG): container finished" podID="ba10e6ec-3b28-462a-b625-a79d63eb5f92" containerID="fb74a1d1398a99d7b226741161cc879ac7da92c408cc2bbce5fdba987d7cd177" exitCode=0 Dec 01 17:41:36 crc kubenswrapper[4868]: I1201 17:41:36.940957 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" event={"ID":"ba10e6ec-3b28-462a-b625-a79d63eb5f92","Type":"ContainerDied","Data":"fb74a1d1398a99d7b226741161cc879ac7da92c408cc2bbce5fdba987d7cd177"} Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.049137 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q7782"] Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.074909 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.120393 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z4f4z"] Dec 01 17:41:37 crc kubenswrapper[4868]: W1201 17:41:37.123848 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5daf33c_c3fa_4f86_a422_93e7afa1afaa.slice/crio-f11a96e3bc572781b287a40771b84921ac365c5adfc5c5ae04545ca0a5395442 WatchSource:0}: Error finding container f11a96e3bc572781b287a40771b84921ac365c5adfc5c5ae04545ca0a5395442: Status 404 returned error can't find the container with id f11a96e3bc572781b287a40771b84921ac365c5adfc5c5ae04545ca0a5395442 Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.148419 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.212854 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 17:41:37 crc kubenswrapper[4868]: W1201 17:41:37.226481 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89b376d8_f72b_4c03_a2f1_3fcf6ff20336.slice/crio-dfebbd1842769f07d04fc12a363f20a4f52b0981cc236bfea6d10f9cd4b52c33 WatchSource:0}: Error finding container dfebbd1842769f07d04fc12a363f20a4f52b0981cc236bfea6d10f9cd4b52c33: Status 404 returned error can't find the container with id dfebbd1842769f07d04fc12a363f20a4f52b0981cc236bfea6d10f9cd4b52c33 Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.305371 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 17:41:37 crc kubenswrapper[4868]: E1201 17:41:37.345136 4868 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 01 17:41:37 crc kubenswrapper[4868]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/6d71bfbf-8024-47e4-860c-c1cb363bb185/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 01 17:41:37 crc kubenswrapper[4868]: > podSandboxID="33dfa299d9ddfdb9a56c6f14a13020d5cb737c8b566c54f5bb6337ff2b0b4230" Dec 01 17:41:37 crc kubenswrapper[4868]: E1201 17:41:37.345684 4868 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 01 17:41:37 crc kubenswrapper[4868]: init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6ftkk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-7ch5d_openstack(6d71bfbf-8024-47e4-860c-c1cb363bb185): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/6d71bfbf-8024-47e4-860c-c1cb363bb185/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 01 17:41:37 crc kubenswrapper[4868]: > logger="UnhandledError" Dec 01 17:41:37 crc kubenswrapper[4868]: E1201 17:41:37.347237 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/6d71bfbf-8024-47e4-860c-c1cb363bb185/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" podUID="6d71bfbf-8024-47e4-860c-c1cb363bb185" Dec 01 17:41:37 crc kubenswrapper[4868]: W1201 17:41:37.421184 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfaa33685_51f4_46ce_bdc7_ef9a2467f482.slice/crio-467c07b3f10d946ef80830b76bb20f59ef7cdcf8f9caf5b24810ca386915c77e WatchSource:0}: Error finding container 467c07b3f10d946ef80830b76bb20f59ef7cdcf8f9caf5b24810ca386915c77e: Status 404 returned error can't find the container with id 467c07b3f10d946ef80830b76bb20f59ef7cdcf8f9caf5b24810ca386915c77e Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.459347 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-vjpd8" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.473898 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.620708 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cec39435-608a-497c-8802-0cc4639599bf-config\") pod \"cec39435-608a-497c-8802-0cc4639599bf\" (UID: \"cec39435-608a-497c-8802-0cc4639599bf\") " Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.620835 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cec39435-608a-497c-8802-0cc4639599bf-dns-svc\") pod \"cec39435-608a-497c-8802-0cc4639599bf\" (UID: \"cec39435-608a-497c-8802-0cc4639599bf\") " Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.620893 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cafaee0-d17b-4e2e-a6c4-67382ce926c9-config\") pod \"7cafaee0-d17b-4e2e-a6c4-67382ce926c9\" (UID: \"7cafaee0-d17b-4e2e-a6c4-67382ce926c9\") " Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.620917 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qnlf\" (UniqueName: \"kubernetes.io/projected/7cafaee0-d17b-4e2e-a6c4-67382ce926c9-kube-api-access-4qnlf\") pod \"7cafaee0-d17b-4e2e-a6c4-67382ce926c9\" (UID: \"7cafaee0-d17b-4e2e-a6c4-67382ce926c9\") " Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.620986 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6j7t7\" (UniqueName: \"kubernetes.io/projected/cec39435-608a-497c-8802-0cc4639599bf-kube-api-access-6j7t7\") pod \"cec39435-608a-497c-8802-0cc4639599bf\" (UID: \"cec39435-608a-497c-8802-0cc4639599bf\") " Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.622465 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cec39435-608a-497c-8802-0cc4639599bf-config" (OuterVolumeSpecName: "config") pod "cec39435-608a-497c-8802-0cc4639599bf" (UID: "cec39435-608a-497c-8802-0cc4639599bf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.622640 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cafaee0-d17b-4e2e-a6c4-67382ce926c9-config" (OuterVolumeSpecName: "config") pod "7cafaee0-d17b-4e2e-a6c4-67382ce926c9" (UID: "7cafaee0-d17b-4e2e-a6c4-67382ce926c9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.622758 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cec39435-608a-497c-8802-0cc4639599bf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cec39435-608a-497c-8802-0cc4639599bf" (UID: "cec39435-608a-497c-8802-0cc4639599bf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.628151 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cafaee0-d17b-4e2e-a6c4-67382ce926c9-kube-api-access-4qnlf" (OuterVolumeSpecName: "kube-api-access-4qnlf") pod "7cafaee0-d17b-4e2e-a6c4-67382ce926c9" (UID: "7cafaee0-d17b-4e2e-a6c4-67382ce926c9"). InnerVolumeSpecName "kube-api-access-4qnlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.628258 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cec39435-608a-497c-8802-0cc4639599bf-kube-api-access-6j7t7" (OuterVolumeSpecName: "kube-api-access-6j7t7") pod "cec39435-608a-497c-8802-0cc4639599bf" (UID: "cec39435-608a-497c-8802-0cc4639599bf"). InnerVolumeSpecName "kube-api-access-6j7t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.722926 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cec39435-608a-497c-8802-0cc4639599bf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.722993 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cafaee0-d17b-4e2e-a6c4-67382ce926c9-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.723004 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qnlf\" (UniqueName: \"kubernetes.io/projected/7cafaee0-d17b-4e2e-a6c4-67382ce926c9-kube-api-access-4qnlf\") on node \"crc\" DevicePath \"\"" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.723017 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6j7t7\" (UniqueName: \"kubernetes.io/projected/cec39435-608a-497c-8802-0cc4639599bf-kube-api-access-6j7t7\") on node \"crc\" DevicePath \"\"" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.723027 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cec39435-608a-497c-8802-0cc4639599bf-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.951538 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-vjpd8" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.951570 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-vjpd8" event={"ID":"7cafaee0-d17b-4e2e-a6c4-67382ce926c9","Type":"ContainerDied","Data":"94248d23982b57cb9b991a48f4a72075dd001a8d497f5fda8f00670f1fa09d83"} Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.952640 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" event={"ID":"cec39435-608a-497c-8802-0cc4639599bf","Type":"ContainerDied","Data":"b1447c6951c2d423dc8c7a8086cccd92c9eba647690079d971d758e6867f71c5"} Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.952695 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6r4nr" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.953537 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"faa33685-51f4-46ce-bdc7-ef9a2467f482","Type":"ContainerStarted","Data":"467c07b3f10d946ef80830b76bb20f59ef7cdcf8f9caf5b24810ca386915c77e"} Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.956432 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" event={"ID":"ba10e6ec-3b28-462a-b625-a79d63eb5f92","Type":"ContainerStarted","Data":"a70f2cb3ad7c12355fb4be946c123ae616452b53e93b86b1971105d87dc0ebdc"} Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.957055 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.959066 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z4f4z" event={"ID":"65d7a832-930e-4103-90f6-dbc5de8c1ece","Type":"ContainerStarted","Data":"e43db2cec30a66855d7e9626971db6c7345efcc1d2a39288ea9ee063d17b8450"} Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.960197 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d5daf33c-c3fa-4f86-a422-93e7afa1afaa","Type":"ContainerStarted","Data":"f11a96e3bc572781b287a40771b84921ac365c5adfc5c5ae04545ca0a5395442"} Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.976342 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcxdx" event={"ID":"a904bedc-9d01-4a00-82c1-6f2826a33d81","Type":"ContainerStarted","Data":"4baa271331fb73640e1874f41c7b8aef42f823f32748017c9b82915b0f40db4a"} Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.986065 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" podStartSLOduration=12.451527875 podStartE2EDuration="28.986038761s" podCreationTimestamp="2025-12-01 17:41:09 +0000 UTC" firstStartedPulling="2025-12-01 17:41:19.587125063 +0000 UTC m=+951.958235474" lastFinishedPulling="2025-12-01 17:41:36.121635949 +0000 UTC m=+968.492746360" observedRunningTime="2025-12-01 17:41:37.979238824 +0000 UTC m=+970.350349245" watchObservedRunningTime="2025-12-01 17:41:37.986038761 +0000 UTC m=+970.357149172" Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.988035 4868 generic.go:334] "Generic (PLEG): container finished" podID="d7e2ea14-2a0a-418c-ab25-4a588d728f5a" containerID="ef5be524361fca9d05a4d37d50616c9ed37d89cce7d86bcaa3672d530fb55536" exitCode=0 Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.988159 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7782" event={"ID":"d7e2ea14-2a0a-418c-ab25-4a588d728f5a","Type":"ContainerDied","Data":"ef5be524361fca9d05a4d37d50616c9ed37d89cce7d86bcaa3672d530fb55536"} Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.988204 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7782" event={"ID":"d7e2ea14-2a0a-418c-ab25-4a588d728f5a","Type":"ContainerStarted","Data":"7f8ec6b6373e6427938cd7a070a5c99df781f7f7c61ce8076b07594113ec91fa"} Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.991389 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cc48650c-5f3d-4c0a-b570-c7eb8615d504","Type":"ContainerStarted","Data":"fc551053630e0c330e7eefd489ccca2202199023d89caf5f5d6184aff7b04753"} Dec 01 17:41:37 crc kubenswrapper[4868]: I1201 17:41:37.999644 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rcxdx" podStartSLOduration=2.128609184 podStartE2EDuration="35.999623205s" podCreationTimestamp="2025-12-01 17:41:02 +0000 UTC" firstStartedPulling="2025-12-01 17:41:03.524970072 +0000 UTC m=+935.896080493" lastFinishedPulling="2025-12-01 17:41:37.395984103 +0000 UTC m=+969.767094514" observedRunningTime="2025-12-01 17:41:37.998043621 +0000 UTC m=+970.369154042" watchObservedRunningTime="2025-12-01 17:41:37.999623205 +0000 UTC m=+970.370733616" Dec 01 17:41:38 crc kubenswrapper[4868]: I1201 17:41:38.001117 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"89b376d8-f72b-4c03-a2f1-3fcf6ff20336","Type":"ContainerStarted","Data":"dfebbd1842769f07d04fc12a363f20a4f52b0981cc236bfea6d10f9cd4b52c33"} Dec 01 17:41:38 crc kubenswrapper[4868]: I1201 17:41:38.086397 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rzgnl"] Dec 01 17:41:38 crc kubenswrapper[4868]: I1201 17:41:38.147402 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6r4nr"] Dec 01 17:41:38 crc kubenswrapper[4868]: I1201 17:41:38.164036 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6r4nr"] Dec 01 17:41:38 crc kubenswrapper[4868]: I1201 17:41:38.208419 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cec39435-608a-497c-8802-0cc4639599bf" path="/var/lib/kubelet/pods/cec39435-608a-497c-8802-0cc4639599bf/volumes" Dec 01 17:41:38 crc kubenswrapper[4868]: I1201 17:41:38.229218 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vjpd8"] Dec 01 17:41:38 crc kubenswrapper[4868]: I1201 17:41:38.239835 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vjpd8"] Dec 01 17:41:40 crc kubenswrapper[4868]: I1201 17:41:40.197770 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cafaee0-d17b-4e2e-a6c4-67382ce926c9" path="/var/lib/kubelet/pods/7cafaee0-d17b-4e2e-a6c4-67382ce926c9/volumes" Dec 01 17:41:41 crc kubenswrapper[4868]: I1201 17:41:41.032628 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rzgnl" event={"ID":"8cca4bfb-cddc-425e-abe2-6debea4aa92a","Type":"ContainerStarted","Data":"6dd56ef14d664df60875598fec2d1cf5797e3759d338297f7a9721019ac3e443"} Dec 01 17:41:42 crc kubenswrapper[4868]: I1201 17:41:42.816066 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rcxdx" Dec 01 17:41:42 crc kubenswrapper[4868]: I1201 17:41:42.817830 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rcxdx" Dec 01 17:41:42 crc kubenswrapper[4868]: I1201 17:41:42.867880 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rcxdx" Dec 01 17:41:43 crc kubenswrapper[4868]: I1201 17:41:43.124490 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rcxdx" Dec 01 17:41:43 crc kubenswrapper[4868]: I1201 17:41:43.249733 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rcxdx"] Dec 01 17:41:43 crc kubenswrapper[4868]: I1201 17:41:43.351364 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ckkq6"] Dec 01 17:41:43 crc kubenswrapper[4868]: I1201 17:41:43.351662 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ckkq6" podUID="901c8b2e-6560-4d33-8233-1f01ed46b2b5" containerName="registry-server" containerID="cri-o://97dc7423675eb5f575f9660a1f97eeeb6a9090124b404f56eef1bd8461f793bf" gracePeriod=2 Dec 01 17:41:47 crc kubenswrapper[4868]: I1201 17:41:44.802159 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" Dec 01 17:41:47 crc kubenswrapper[4868]: I1201 17:41:44.871416 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7ch5d"] Dec 01 17:41:47 crc kubenswrapper[4868]: I1201 17:41:45.069247 4868 generic.go:334] "Generic (PLEG): container finished" podID="901c8b2e-6560-4d33-8233-1f01ed46b2b5" containerID="97dc7423675eb5f575f9660a1f97eeeb6a9090124b404f56eef1bd8461f793bf" exitCode=0 Dec 01 17:41:47 crc kubenswrapper[4868]: I1201 17:41:45.069465 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ckkq6" event={"ID":"901c8b2e-6560-4d33-8233-1f01ed46b2b5","Type":"ContainerDied","Data":"97dc7423675eb5f575f9660a1f97eeeb6a9090124b404f56eef1bd8461f793bf"} Dec 01 17:41:51 crc kubenswrapper[4868]: I1201 17:41:51.585356 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" Dec 01 17:41:51 crc kubenswrapper[4868]: I1201 17:41:51.713448 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d71bfbf-8024-47e4-860c-c1cb363bb185-dns-svc\") pod \"6d71bfbf-8024-47e4-860c-c1cb363bb185\" (UID: \"6d71bfbf-8024-47e4-860c-c1cb363bb185\") " Dec 01 17:41:51 crc kubenswrapper[4868]: I1201 17:41:51.714157 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d71bfbf-8024-47e4-860c-c1cb363bb185-config\") pod \"6d71bfbf-8024-47e4-860c-c1cb363bb185\" (UID: \"6d71bfbf-8024-47e4-860c-c1cb363bb185\") " Dec 01 17:41:51 crc kubenswrapper[4868]: I1201 17:41:51.714229 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ftkk\" (UniqueName: \"kubernetes.io/projected/6d71bfbf-8024-47e4-860c-c1cb363bb185-kube-api-access-6ftkk\") pod \"6d71bfbf-8024-47e4-860c-c1cb363bb185\" (UID: \"6d71bfbf-8024-47e4-860c-c1cb363bb185\") " Dec 01 17:41:51 crc kubenswrapper[4868]: I1201 17:41:51.730527 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d71bfbf-8024-47e4-860c-c1cb363bb185-kube-api-access-6ftkk" (OuterVolumeSpecName: "kube-api-access-6ftkk") pod "6d71bfbf-8024-47e4-860c-c1cb363bb185" (UID: "6d71bfbf-8024-47e4-860c-c1cb363bb185"). InnerVolumeSpecName "kube-api-access-6ftkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:41:51 crc kubenswrapper[4868]: I1201 17:41:51.742303 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d71bfbf-8024-47e4-860c-c1cb363bb185-config" (OuterVolumeSpecName: "config") pod "6d71bfbf-8024-47e4-860c-c1cb363bb185" (UID: "6d71bfbf-8024-47e4-860c-c1cb363bb185"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:41:51 crc kubenswrapper[4868]: I1201 17:41:51.748147 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d71bfbf-8024-47e4-860c-c1cb363bb185-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6d71bfbf-8024-47e4-860c-c1cb363bb185" (UID: "6d71bfbf-8024-47e4-860c-c1cb363bb185"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:41:51 crc kubenswrapper[4868]: I1201 17:41:51.816929 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d71bfbf-8024-47e4-860c-c1cb363bb185-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:41:51 crc kubenswrapper[4868]: I1201 17:41:51.816997 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ftkk\" (UniqueName: \"kubernetes.io/projected/6d71bfbf-8024-47e4-860c-c1cb363bb185-kube-api-access-6ftkk\") on node \"crc\" DevicePath \"\"" Dec 01 17:41:51 crc kubenswrapper[4868]: I1201 17:41:51.817019 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d71bfbf-8024-47e4-860c-c1cb363bb185-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:41:51 crc kubenswrapper[4868]: E1201 17:41:51.926185 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified" Dec 01 17:41:51 crc kubenswrapper[4868]: E1201 17:41:51.926417 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-nb,Image:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n549h647hf6h678hcdh58ch5cfh65chd8h79h68h666h5c8h69hd4h555h5dfh78hc6h57bh65fh644h7fh5cdh5dch584hb8h75h66dh64dh5f4h5c7q,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-nb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-75sk2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(faa33685-51f4-46ce-bdc7-ef9a2467f482): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 17:41:52 crc kubenswrapper[4868]: I1201 17:41:52.123752 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" event={"ID":"6d71bfbf-8024-47e4-860c-c1cb363bb185","Type":"ContainerDied","Data":"33dfa299d9ddfdb9a56c6f14a13020d5cb737c8b566c54f5bb6337ff2b0b4230"} Dec 01 17:41:52 crc kubenswrapper[4868]: I1201 17:41:52.123799 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7ch5d" Dec 01 17:41:52 crc kubenswrapper[4868]: I1201 17:41:52.192286 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7ch5d"] Dec 01 17:41:52 crc kubenswrapper[4868]: I1201 17:41:52.201657 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7ch5d"] Dec 01 17:41:52 crc kubenswrapper[4868]: I1201 17:41:52.462391 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:41:52 crc kubenswrapper[4868]: I1201 17:41:52.630072 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/901c8b2e-6560-4d33-8233-1f01ed46b2b5-utilities\") pod \"901c8b2e-6560-4d33-8233-1f01ed46b2b5\" (UID: \"901c8b2e-6560-4d33-8233-1f01ed46b2b5\") " Dec 01 17:41:52 crc kubenswrapper[4868]: I1201 17:41:52.630230 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/901c8b2e-6560-4d33-8233-1f01ed46b2b5-catalog-content\") pod \"901c8b2e-6560-4d33-8233-1f01ed46b2b5\" (UID: \"901c8b2e-6560-4d33-8233-1f01ed46b2b5\") " Dec 01 17:41:52 crc kubenswrapper[4868]: I1201 17:41:52.630381 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rz8tr\" (UniqueName: \"kubernetes.io/projected/901c8b2e-6560-4d33-8233-1f01ed46b2b5-kube-api-access-rz8tr\") pod \"901c8b2e-6560-4d33-8233-1f01ed46b2b5\" (UID: \"901c8b2e-6560-4d33-8233-1f01ed46b2b5\") " Dec 01 17:41:52 crc kubenswrapper[4868]: I1201 17:41:52.632104 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/901c8b2e-6560-4d33-8233-1f01ed46b2b5-utilities" (OuterVolumeSpecName: "utilities") pod "901c8b2e-6560-4d33-8233-1f01ed46b2b5" (UID: "901c8b2e-6560-4d33-8233-1f01ed46b2b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:41:52 crc kubenswrapper[4868]: I1201 17:41:52.636847 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/901c8b2e-6560-4d33-8233-1f01ed46b2b5-kube-api-access-rz8tr" (OuterVolumeSpecName: "kube-api-access-rz8tr") pod "901c8b2e-6560-4d33-8233-1f01ed46b2b5" (UID: "901c8b2e-6560-4d33-8233-1f01ed46b2b5"). InnerVolumeSpecName "kube-api-access-rz8tr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:41:52 crc kubenswrapper[4868]: I1201 17:41:52.732630 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/901c8b2e-6560-4d33-8233-1f01ed46b2b5-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:41:52 crc kubenswrapper[4868]: I1201 17:41:52.732671 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rz8tr\" (UniqueName: \"kubernetes.io/projected/901c8b2e-6560-4d33-8233-1f01ed46b2b5-kube-api-access-rz8tr\") on node \"crc\" DevicePath \"\"" Dec 01 17:41:52 crc kubenswrapper[4868]: I1201 17:41:52.732682 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/901c8b2e-6560-4d33-8233-1f01ed46b2b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "901c8b2e-6560-4d33-8233-1f01ed46b2b5" (UID: "901c8b2e-6560-4d33-8233-1f01ed46b2b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:41:52 crc kubenswrapper[4868]: E1201 17:41:52.749346 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified" Dec 01 17:41:52 crc kubenswrapper[4868]: E1201 17:41:52.749551 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-sb,Image:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n57bhd4h689h7bh555h9bh667h575hf5h648h55fh658h57h68fh5ch688hbh5f7h5b6h67h7h5bdh5dfh57hcch68dh58fhffh549h64ch68bh5ffq,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-sb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b4xgd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(89b376d8-f72b-4c03-a2f1-3fcf6ff20336): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 17:41:52 crc kubenswrapper[4868]: I1201 17:41:52.834226 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/901c8b2e-6560-4d33-8233-1f01ed46b2b5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:41:53 crc kubenswrapper[4868]: I1201 17:41:53.144304 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ckkq6" event={"ID":"901c8b2e-6560-4d33-8233-1f01ed46b2b5","Type":"ContainerDied","Data":"86678aec97a45d02dff866087ac60031e47bf279656436581fa8f819db096f9b"} Dec 01 17:41:53 crc kubenswrapper[4868]: I1201 17:41:53.144708 4868 scope.go:117] "RemoveContainer" containerID="97dc7423675eb5f575f9660a1f97eeeb6a9090124b404f56eef1bd8461f793bf" Dec 01 17:41:53 crc kubenswrapper[4868]: I1201 17:41:53.144639 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ckkq6" Dec 01 17:41:53 crc kubenswrapper[4868]: I1201 17:41:53.191988 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ckkq6"] Dec 01 17:41:53 crc kubenswrapper[4868]: I1201 17:41:53.204545 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ckkq6"] Dec 01 17:41:53 crc kubenswrapper[4868]: I1201 17:41:53.322555 4868 scope.go:117] "RemoveContainer" containerID="8f2fe03da4a9272ef15092e2794d94c916a1c8cdd19cc20da81456ff942d7ae3" Dec 01 17:41:53 crc kubenswrapper[4868]: E1201 17:41:53.407467 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 01 17:41:53 crc kubenswrapper[4868]: E1201 17:41:53.407547 4868 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 01 17:41:53 crc kubenswrapper[4868]: E1201 17:41:53.407957 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kktwt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(d5daf33c-c3fa-4f86-a422-93e7afa1afaa): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled" logger="UnhandledError" Dec 01 17:41:53 crc kubenswrapper[4868]: E1201 17:41:53.409210 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="d5daf33c-c3fa-4f86-a422-93e7afa1afaa" Dec 01 17:41:53 crc kubenswrapper[4868]: I1201 17:41:53.449418 4868 scope.go:117] "RemoveContainer" containerID="fb4a9edadc208dba4a5b062181a3005ed052eb60cf5beed34bd882c777762e02" Dec 01 17:41:54 crc kubenswrapper[4868]: I1201 17:41:54.156054 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z4f4z" event={"ID":"65d7a832-930e-4103-90f6-dbc5de8c1ece","Type":"ContainerStarted","Data":"cd0e318ff00bb3dadf486db71fa799f635c8a115e836584a3b53d9a4b1e4361b"} Dec 01 17:41:54 crc kubenswrapper[4868]: I1201 17:41:54.157141 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-z4f4z" Dec 01 17:41:54 crc kubenswrapper[4868]: I1201 17:41:54.160613 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e","Type":"ContainerStarted","Data":"77fd755bd9c73c2759d76d93b79072dd8b91cfb634b95eede967ff767f254d8a"} Dec 01 17:41:54 crc kubenswrapper[4868]: I1201 17:41:54.162274 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 01 17:41:54 crc kubenswrapper[4868]: I1201 17:41:54.170303 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rzgnl" event={"ID":"8cca4bfb-cddc-425e-abe2-6debea4aa92a","Type":"ContainerStarted","Data":"2bbcf21272009e0e5a6bd85a6c19fcdeb16b1a3cf95cc49bed822ce422ad3b13"} Dec 01 17:41:54 crc kubenswrapper[4868]: I1201 17:41:54.182572 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d71bfbf-8024-47e4-860c-c1cb363bb185" path="/var/lib/kubelet/pods/6d71bfbf-8024-47e4-860c-c1cb363bb185/volumes" Dec 01 17:41:54 crc kubenswrapper[4868]: I1201 17:41:54.183316 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="901c8b2e-6560-4d33-8233-1f01ed46b2b5" path="/var/lib/kubelet/pods/901c8b2e-6560-4d33-8233-1f01ed46b2b5/volumes" Dec 01 17:41:54 crc kubenswrapper[4868]: I1201 17:41:54.185082 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7782" event={"ID":"d7e2ea14-2a0a-418c-ab25-4a588d728f5a","Type":"ContainerStarted","Data":"f7ec7b64cd4b149b3a6a9f00756bb77a57db9aef036b0f79ba14c7740b234b56"} Dec 01 17:41:54 crc kubenswrapper[4868]: I1201 17:41:54.187446 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"68f3c214-92ad-43c0-b06a-33c6f1f90cfb","Type":"ContainerStarted","Data":"162934f9653c1a1ce34795b0fa3f5db5f20a6585ca65ea78720b605ce77f9d1f"} Dec 01 17:41:54 crc kubenswrapper[4868]: I1201 17:41:54.192247 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cc48650c-5f3d-4c0a-b570-c7eb8615d504","Type":"ContainerStarted","Data":"f55c3aec3a14c97070eb51b67046f81026ee9fd7fcd36afaa314e368e7ecaece"} Dec 01 17:41:54 crc kubenswrapper[4868]: E1201 17:41:54.193123 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="d5daf33c-c3fa-4f86-a422-93e7afa1afaa" Dec 01 17:41:54 crc kubenswrapper[4868]: I1201 17:41:54.196994 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-z4f4z" podStartSLOduration=20.37045236 podStartE2EDuration="36.196966455s" podCreationTimestamp="2025-12-01 17:41:18 +0000 UTC" firstStartedPulling="2025-12-01 17:41:37.17122767 +0000 UTC m=+969.542338081" lastFinishedPulling="2025-12-01 17:41:52.997741765 +0000 UTC m=+985.368852176" observedRunningTime="2025-12-01 17:41:54.185449992 +0000 UTC m=+986.556560403" watchObservedRunningTime="2025-12-01 17:41:54.196966455 +0000 UTC m=+986.568076866" Dec 01 17:41:54 crc kubenswrapper[4868]: I1201 17:41:54.212400 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=25.710585616 podStartE2EDuration="41.212382908s" podCreationTimestamp="2025-12-01 17:41:13 +0000 UTC" firstStartedPulling="2025-12-01 17:41:36.574686394 +0000 UTC m=+968.945796805" lastFinishedPulling="2025-12-01 17:41:52.076483686 +0000 UTC m=+984.447594097" observedRunningTime="2025-12-01 17:41:54.209830553 +0000 UTC m=+986.580940964" watchObservedRunningTime="2025-12-01 17:41:54.212382908 +0000 UTC m=+986.583493319" Dec 01 17:41:55 crc kubenswrapper[4868]: I1201 17:41:55.214345 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3ff4f488-06c8-4ceb-848a-4ba68b334752","Type":"ContainerStarted","Data":"bbc47c78ecea10373d709f8016370bad805694e3d2800bf35760bd08eb36dc39"} Dec 01 17:41:55 crc kubenswrapper[4868]: I1201 17:41:55.237453 4868 generic.go:334] "Generic (PLEG): container finished" podID="8cca4bfb-cddc-425e-abe2-6debea4aa92a" containerID="2bbcf21272009e0e5a6bd85a6c19fcdeb16b1a3cf95cc49bed822ce422ad3b13" exitCode=0 Dec 01 17:41:55 crc kubenswrapper[4868]: I1201 17:41:55.237542 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rzgnl" event={"ID":"8cca4bfb-cddc-425e-abe2-6debea4aa92a","Type":"ContainerDied","Data":"2bbcf21272009e0e5a6bd85a6c19fcdeb16b1a3cf95cc49bed822ce422ad3b13"} Dec 01 17:41:55 crc kubenswrapper[4868]: I1201 17:41:55.248390 4868 generic.go:334] "Generic (PLEG): container finished" podID="d7e2ea14-2a0a-418c-ab25-4a588d728f5a" containerID="f7ec7b64cd4b149b3a6a9f00756bb77a57db9aef036b0f79ba14c7740b234b56" exitCode=0 Dec 01 17:41:55 crc kubenswrapper[4868]: I1201 17:41:55.248700 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7782" event={"ID":"d7e2ea14-2a0a-418c-ab25-4a588d728f5a","Type":"ContainerDied","Data":"f7ec7b64cd4b149b3a6a9f00756bb77a57db9aef036b0f79ba14c7740b234b56"} Dec 01 17:41:55 crc kubenswrapper[4868]: I1201 17:41:55.253876 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"70aee7bf-d819-4355-a98c-4a106f3e6f16","Type":"ContainerStarted","Data":"571a7ff19cbb3694810ea83e478a19bfa550fa8fb19697462b8c615086834487"} Dec 01 17:41:55 crc kubenswrapper[4868]: I1201 17:41:55.904546 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:41:55 crc kubenswrapper[4868]: I1201 17:41:55.905170 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:41:57 crc kubenswrapper[4868]: E1201 17:41:57.000699 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="faa33685-51f4-46ce-bdc7-ef9a2467f482" Dec 01 17:41:57 crc kubenswrapper[4868]: E1201 17:41:57.003114 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="89b376d8-f72b-4c03-a2f1-3fcf6ff20336" Dec 01 17:41:57 crc kubenswrapper[4868]: I1201 17:41:57.273162 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"89b376d8-f72b-4c03-a2f1-3fcf6ff20336","Type":"ContainerStarted","Data":"771129064c34646021f771e6598ec0cb3f192463afc021625523400af544cf1a"} Dec 01 17:41:57 crc kubenswrapper[4868]: E1201 17:41:57.275322 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="89b376d8-f72b-4c03-a2f1-3fcf6ff20336" Dec 01 17:41:57 crc kubenswrapper[4868]: I1201 17:41:57.276791 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rzgnl" event={"ID":"8cca4bfb-cddc-425e-abe2-6debea4aa92a","Type":"ContainerStarted","Data":"5cfaa652bb6e139b509248543d7ac1708090f7a54546dcabaa6704602abe770d"} Dec 01 17:41:57 crc kubenswrapper[4868]: I1201 17:41:57.282163 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7782" event={"ID":"d7e2ea14-2a0a-418c-ab25-4a588d728f5a","Type":"ContainerStarted","Data":"25ef767abfd397a3463059df500f412275ac41704ca813fd6f2f7caf674c01d9"} Dec 01 17:41:57 crc kubenswrapper[4868]: I1201 17:41:57.286274 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"faa33685-51f4-46ce-bdc7-ef9a2467f482","Type":"ContainerStarted","Data":"dad2dbaf60813e7f7290722a0c11e08cf81b1542e7ba361f77835677cc01bc1e"} Dec 01 17:41:57 crc kubenswrapper[4868]: E1201 17:41:57.288409 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="faa33685-51f4-46ce-bdc7-ef9a2467f482" Dec 01 17:41:57 crc kubenswrapper[4868]: I1201 17:41:57.332000 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q7782" podStartSLOduration=9.28688461 podStartE2EDuration="26.331969208s" podCreationTimestamp="2025-12-01 17:41:31 +0000 UTC" firstStartedPulling="2025-12-01 17:41:39.564309593 +0000 UTC m=+971.935420004" lastFinishedPulling="2025-12-01 17:41:56.609394191 +0000 UTC m=+988.980504602" observedRunningTime="2025-12-01 17:41:57.326387116 +0000 UTC m=+989.697497537" watchObservedRunningTime="2025-12-01 17:41:57.331969208 +0000 UTC m=+989.703079629" Dec 01 17:41:58 crc kubenswrapper[4868]: I1201 17:41:58.297697 4868 generic.go:334] "Generic (PLEG): container finished" podID="68f3c214-92ad-43c0-b06a-33c6f1f90cfb" containerID="162934f9653c1a1ce34795b0fa3f5db5f20a6585ca65ea78720b605ce77f9d1f" exitCode=0 Dec 01 17:41:58 crc kubenswrapper[4868]: I1201 17:41:58.297842 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"68f3c214-92ad-43c0-b06a-33c6f1f90cfb","Type":"ContainerDied","Data":"162934f9653c1a1ce34795b0fa3f5db5f20a6585ca65ea78720b605ce77f9d1f"} Dec 01 17:41:58 crc kubenswrapper[4868]: I1201 17:41:58.302806 4868 generic.go:334] "Generic (PLEG): container finished" podID="cc48650c-5f3d-4c0a-b570-c7eb8615d504" containerID="f55c3aec3a14c97070eb51b67046f81026ee9fd7fcd36afaa314e368e7ecaece" exitCode=0 Dec 01 17:41:58 crc kubenswrapper[4868]: I1201 17:41:58.302895 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cc48650c-5f3d-4c0a-b570-c7eb8615d504","Type":"ContainerDied","Data":"f55c3aec3a14c97070eb51b67046f81026ee9fd7fcd36afaa314e368e7ecaece"} Dec 01 17:41:58 crc kubenswrapper[4868]: I1201 17:41:58.307549 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rzgnl" event={"ID":"8cca4bfb-cddc-425e-abe2-6debea4aa92a","Type":"ContainerStarted","Data":"34593bb7551c9fa0bc6e114f06a021572674dfb985f438a0c1144afd024a03c1"} Dec 01 17:41:58 crc kubenswrapper[4868]: I1201 17:41:58.307692 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:58 crc kubenswrapper[4868]: I1201 17:41:58.308050 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:41:58 crc kubenswrapper[4868]: E1201 17:41:58.309186 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="89b376d8-f72b-4c03-a2f1-3fcf6ff20336" Dec 01 17:41:58 crc kubenswrapper[4868]: E1201 17:41:58.309694 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="faa33685-51f4-46ce-bdc7-ef9a2467f482" Dec 01 17:41:58 crc kubenswrapper[4868]: I1201 17:41:58.442683 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-rzgnl" podStartSLOduration=28.135151559 podStartE2EDuration="40.442662886s" podCreationTimestamp="2025-12-01 17:41:18 +0000 UTC" firstStartedPulling="2025-12-01 17:41:40.60497262 +0000 UTC m=+972.976083031" lastFinishedPulling="2025-12-01 17:41:52.912483947 +0000 UTC m=+985.283594358" observedRunningTime="2025-12-01 17:41:58.439064425 +0000 UTC m=+990.810174836" watchObservedRunningTime="2025-12-01 17:41:58.442662886 +0000 UTC m=+990.813773297" Dec 01 17:41:58 crc kubenswrapper[4868]: I1201 17:41:58.683457 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 01 17:41:59 crc kubenswrapper[4868]: I1201 17:41:59.322356 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cc48650c-5f3d-4c0a-b570-c7eb8615d504","Type":"ContainerStarted","Data":"dc8e6d680f76954297e800753b4bc475d88a08a527cb0fef22377fdb780d5201"} Dec 01 17:41:59 crc kubenswrapper[4868]: I1201 17:41:59.328444 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"68f3c214-92ad-43c0-b06a-33c6f1f90cfb","Type":"ContainerStarted","Data":"66000af62e51102ed6ea72bd9f97d9f8db762f5617fef15ca1e56eb736b3f79c"} Dec 01 17:41:59 crc kubenswrapper[4868]: I1201 17:41:59.378136 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=31.45679021 podStartE2EDuration="47.378109998s" podCreationTimestamp="2025-12-01 17:41:12 +0000 UTC" firstStartedPulling="2025-12-01 17:41:37.160968748 +0000 UTC m=+969.532079159" lastFinishedPulling="2025-12-01 17:41:53.082288536 +0000 UTC m=+985.453398947" observedRunningTime="2025-12-01 17:41:59.360764787 +0000 UTC m=+991.731875208" watchObservedRunningTime="2025-12-01 17:41:59.378109998 +0000 UTC m=+991.749220409" Dec 01 17:42:01 crc kubenswrapper[4868]: I1201 17:42:01.833079 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:42:01 crc kubenswrapper[4868]: I1201 17:42:01.833646 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:42:01 crc kubenswrapper[4868]: I1201 17:42:01.891781 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:42:01 crc kubenswrapper[4868]: I1201 17:42:01.917724 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=36.067575534 podStartE2EDuration="51.917699548s" podCreationTimestamp="2025-12-01 17:41:10 +0000 UTC" firstStartedPulling="2025-12-01 17:41:36.582653603 +0000 UTC m=+968.953764014" lastFinishedPulling="2025-12-01 17:41:52.432777617 +0000 UTC m=+984.803888028" observedRunningTime="2025-12-01 17:41:59.425807051 +0000 UTC m=+991.796917462" watchObservedRunningTime="2025-12-01 17:42:01.917699548 +0000 UTC m=+994.288809959" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.403629 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-87c94"] Dec 01 17:42:02 crc kubenswrapper[4868]: E1201 17:42:02.404606 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="901c8b2e-6560-4d33-8233-1f01ed46b2b5" containerName="extract-utilities" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.404634 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="901c8b2e-6560-4d33-8233-1f01ed46b2b5" containerName="extract-utilities" Dec 01 17:42:02 crc kubenswrapper[4868]: E1201 17:42:02.404674 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="901c8b2e-6560-4d33-8233-1f01ed46b2b5" containerName="extract-content" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.404684 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="901c8b2e-6560-4d33-8233-1f01ed46b2b5" containerName="extract-content" Dec 01 17:42:02 crc kubenswrapper[4868]: E1201 17:42:02.404714 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="901c8b2e-6560-4d33-8233-1f01ed46b2b5" containerName="registry-server" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.404725 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="901c8b2e-6560-4d33-8233-1f01ed46b2b5" containerName="registry-server" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.404959 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="901c8b2e-6560-4d33-8233-1f01ed46b2b5" containerName="registry-server" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.405656 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.422889 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.423912 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.424775 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.441657 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-87c94"] Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.444295 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.520275 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-9bttc"] Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.521807 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.527504 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.540476 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q7782"] Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.552397 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpv24\" (UniqueName: \"kubernetes.io/projected/7328948b-0a8c-45d3-91b1-c8317bc12499-kube-api-access-gpv24\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.552532 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7328948b-0a8c-45d3-91b1-c8317bc12499-ovn-rundir\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.552563 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7328948b-0a8c-45d3-91b1-c8317bc12499-ovs-rundir\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.552666 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7328948b-0a8c-45d3-91b1-c8317bc12499-config\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.552706 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7328948b-0a8c-45d3-91b1-c8317bc12499-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.552735 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7328948b-0a8c-45d3-91b1-c8317bc12499-combined-ca-bundle\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.602433 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-9bttc"] Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.654143 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7328948b-0a8c-45d3-91b1-c8317bc12499-ovn-rundir\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.654207 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlzxx\" (UniqueName: \"kubernetes.io/projected/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-kube-api-access-jlzxx\") pod \"dnsmasq-dns-6bc7876d45-9bttc\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.654229 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-9bttc\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.654251 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7328948b-0a8c-45d3-91b1-c8317bc12499-ovs-rundir\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.654327 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7328948b-0a8c-45d3-91b1-c8317bc12499-config\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.654355 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7328948b-0a8c-45d3-91b1-c8317bc12499-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.654504 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-config\") pod \"dnsmasq-dns-6bc7876d45-9bttc\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.654626 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7328948b-0a8c-45d3-91b1-c8317bc12499-combined-ca-bundle\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.654703 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-9bttc\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.654769 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpv24\" (UniqueName: \"kubernetes.io/projected/7328948b-0a8c-45d3-91b1-c8317bc12499-kube-api-access-gpv24\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.654889 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7328948b-0a8c-45d3-91b1-c8317bc12499-ovn-rundir\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.656713 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7328948b-0a8c-45d3-91b1-c8317bc12499-ovs-rundir\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.657906 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7328948b-0a8c-45d3-91b1-c8317bc12499-config\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.690407 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7328948b-0a8c-45d3-91b1-c8317bc12499-combined-ca-bundle\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.690918 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7328948b-0a8c-45d3-91b1-c8317bc12499-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.694633 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpv24\" (UniqueName: \"kubernetes.io/projected/7328948b-0a8c-45d3-91b1-c8317bc12499-kube-api-access-gpv24\") pod \"ovn-controller-metrics-87c94\" (UID: \"7328948b-0a8c-45d3-91b1-c8317bc12499\") " pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.755914 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-87c94" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.758346 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlzxx\" (UniqueName: \"kubernetes.io/projected/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-kube-api-access-jlzxx\") pod \"dnsmasq-dns-6bc7876d45-9bttc\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.758413 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-9bttc\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.758519 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-config\") pod \"dnsmasq-dns-6bc7876d45-9bttc\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.758569 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-9bttc\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.759887 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-9bttc\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.759913 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-config\") pod \"dnsmasq-dns-6bc7876d45-9bttc\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.759922 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-9bttc\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.769758 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-9bttc"] Dec 01 17:42:02 crc kubenswrapper[4868]: E1201 17:42:02.771348 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-jlzxx], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" podUID="edf1d58a-829d-4acc-be67-7c76bc6ed0b9" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.788876 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlzxx\" (UniqueName: \"kubernetes.io/projected/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-kube-api-access-jlzxx\") pod \"dnsmasq-dns-6bc7876d45-9bttc\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.798896 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-b6k5w"] Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.800535 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.808435 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.866903 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-b6k5w"] Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.962461 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p8mx\" (UniqueName: \"kubernetes.io/projected/62062537-295a-4cd8-96e6-1b9961bbfa02-kube-api-access-5p8mx\") pod \"dnsmasq-dns-8554648995-b6k5w\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.962587 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-b6k5w\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.962711 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-b6k5w\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.962771 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-config\") pod \"dnsmasq-dns-8554648995-b6k5w\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:02 crc kubenswrapper[4868]: I1201 17:42:02.962805 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-dns-svc\") pod \"dnsmasq-dns-8554648995-b6k5w\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.073998 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-b6k5w\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.074060 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-config\") pod \"dnsmasq-dns-8554648995-b6k5w\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.074094 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-dns-svc\") pod \"dnsmasq-dns-8554648995-b6k5w\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.074139 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p8mx\" (UniqueName: \"kubernetes.io/projected/62062537-295a-4cd8-96e6-1b9961bbfa02-kube-api-access-5p8mx\") pod \"dnsmasq-dns-8554648995-b6k5w\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.074194 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-b6k5w\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.075152 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-b6k5w\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.075971 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-b6k5w\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.076466 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-config\") pod \"dnsmasq-dns-8554648995-b6k5w\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.090021 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-dns-svc\") pod \"dnsmasq-dns-8554648995-b6k5w\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.108971 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p8mx\" (UniqueName: \"kubernetes.io/projected/62062537-295a-4cd8-96e6-1b9961bbfa02-kube-api-access-5p8mx\") pod \"dnsmasq-dns-8554648995-b6k5w\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.194366 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.241610 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-87c94"] Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.364977 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-87c94" event={"ID":"7328948b-0a8c-45d3-91b1-c8317bc12499","Type":"ContainerStarted","Data":"6781a7801a36815ab3d39bdde0ad308e79556b8c8468810b6a72d3ede5d4007c"} Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.365470 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.382394 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.482727 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-dns-svc\") pod \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.483224 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-config\") pod \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.483296 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlzxx\" (UniqueName: \"kubernetes.io/projected/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-kube-api-access-jlzxx\") pod \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.483331 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-ovsdbserver-sb\") pod \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\" (UID: \"edf1d58a-829d-4acc-be67-7c76bc6ed0b9\") " Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.483910 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-config" (OuterVolumeSpecName: "config") pod "edf1d58a-829d-4acc-be67-7c76bc6ed0b9" (UID: "edf1d58a-829d-4acc-be67-7c76bc6ed0b9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.484128 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "edf1d58a-829d-4acc-be67-7c76bc6ed0b9" (UID: "edf1d58a-829d-4acc-be67-7c76bc6ed0b9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.484258 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "edf1d58a-829d-4acc-be67-7c76bc6ed0b9" (UID: "edf1d58a-829d-4acc-be67-7c76bc6ed0b9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.485341 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.485395 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.485414 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.492198 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-kube-api-access-jlzxx" (OuterVolumeSpecName: "kube-api-access-jlzxx") pod "edf1d58a-829d-4acc-be67-7c76bc6ed0b9" (UID: "edf1d58a-829d-4acc-be67-7c76bc6ed0b9"). InnerVolumeSpecName "kube-api-access-jlzxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.587881 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlzxx\" (UniqueName: \"kubernetes.io/projected/edf1d58a-829d-4acc-be67-7c76bc6ed0b9-kube-api-access-jlzxx\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.663757 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.664341 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 01 17:42:03 crc kubenswrapper[4868]: I1201 17:42:03.755365 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-b6k5w"] Dec 01 17:42:03 crc kubenswrapper[4868]: W1201 17:42:03.762914 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62062537_295a_4cd8_96e6_1b9961bbfa02.slice/crio-74b0ea2f951729d12600025ccc7396ab6618936851d4f6e74f9ca250027dac7e WatchSource:0}: Error finding container 74b0ea2f951729d12600025ccc7396ab6618936851d4f6e74f9ca250027dac7e: Status 404 returned error can't find the container with id 74b0ea2f951729d12600025ccc7396ab6618936851d4f6e74f9ca250027dac7e Dec 01 17:42:04 crc kubenswrapper[4868]: I1201 17:42:04.373718 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-87c94" event={"ID":"7328948b-0a8c-45d3-91b1-c8317bc12499","Type":"ContainerStarted","Data":"d04517e7c0a2d0c45b9c5f8d9409b4aa2da8128a37789e40fbdbb00c22a0d622"} Dec 01 17:42:04 crc kubenswrapper[4868]: I1201 17:42:04.376379 4868 generic.go:334] "Generic (PLEG): container finished" podID="62062537-295a-4cd8-96e6-1b9961bbfa02" containerID="fd35ed0f45997750f2a7b4f15850a6c3b70d2167cddc8f5830248f28cb07140d" exitCode=0 Dec 01 17:42:04 crc kubenswrapper[4868]: I1201 17:42:04.376490 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-b6k5w" event={"ID":"62062537-295a-4cd8-96e6-1b9961bbfa02","Type":"ContainerDied","Data":"fd35ed0f45997750f2a7b4f15850a6c3b70d2167cddc8f5830248f28cb07140d"} Dec 01 17:42:04 crc kubenswrapper[4868]: I1201 17:42:04.376538 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-9bttc" Dec 01 17:42:04 crc kubenswrapper[4868]: I1201 17:42:04.376549 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-b6k5w" event={"ID":"62062537-295a-4cd8-96e6-1b9961bbfa02","Type":"ContainerStarted","Data":"74b0ea2f951729d12600025ccc7396ab6618936851d4f6e74f9ca250027dac7e"} Dec 01 17:42:04 crc kubenswrapper[4868]: I1201 17:42:04.376644 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q7782" podUID="d7e2ea14-2a0a-418c-ab25-4a588d728f5a" containerName="registry-server" containerID="cri-o://25ef767abfd397a3463059df500f412275ac41704ca813fd6f2f7caf674c01d9" gracePeriod=2 Dec 01 17:42:04 crc kubenswrapper[4868]: I1201 17:42:04.396233 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-87c94" podStartSLOduration=2.396213833 podStartE2EDuration="2.396213833s" podCreationTimestamp="2025-12-01 17:42:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:42:04.395621398 +0000 UTC m=+996.766731799" watchObservedRunningTime="2025-12-01 17:42:04.396213833 +0000 UTC m=+996.767324244" Dec 01 17:42:04 crc kubenswrapper[4868]: I1201 17:42:04.480786 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-9bttc"] Dec 01 17:42:04 crc kubenswrapper[4868]: I1201 17:42:04.492769 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-9bttc"] Dec 01 17:42:04 crc kubenswrapper[4868]: I1201 17:42:04.770434 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 01 17:42:04 crc kubenswrapper[4868]: I1201 17:42:04.872127 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 01 17:42:04 crc kubenswrapper[4868]: I1201 17:42:04.916039 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.015671 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-catalog-content\") pod \"d7e2ea14-2a0a-418c-ab25-4a588d728f5a\" (UID: \"d7e2ea14-2a0a-418c-ab25-4a588d728f5a\") " Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.015827 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktplj\" (UniqueName: \"kubernetes.io/projected/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-kube-api-access-ktplj\") pod \"d7e2ea14-2a0a-418c-ab25-4a588d728f5a\" (UID: \"d7e2ea14-2a0a-418c-ab25-4a588d728f5a\") " Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.015959 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-utilities\") pod \"d7e2ea14-2a0a-418c-ab25-4a588d728f5a\" (UID: \"d7e2ea14-2a0a-418c-ab25-4a588d728f5a\") " Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.017158 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-utilities" (OuterVolumeSpecName: "utilities") pod "d7e2ea14-2a0a-418c-ab25-4a588d728f5a" (UID: "d7e2ea14-2a0a-418c-ab25-4a588d728f5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.035562 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-kube-api-access-ktplj" (OuterVolumeSpecName: "kube-api-access-ktplj") pod "d7e2ea14-2a0a-418c-ab25-4a588d728f5a" (UID: "d7e2ea14-2a0a-418c-ab25-4a588d728f5a"). InnerVolumeSpecName "kube-api-access-ktplj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.057803 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7e2ea14-2a0a-418c-ab25-4a588d728f5a" (UID: "d7e2ea14-2a0a-418c-ab25-4a588d728f5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.118685 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.118725 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktplj\" (UniqueName: \"kubernetes.io/projected/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-kube-api-access-ktplj\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.118736 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7e2ea14-2a0a-418c-ab25-4a588d728f5a-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.386136 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-b6k5w" event={"ID":"62062537-295a-4cd8-96e6-1b9961bbfa02","Type":"ContainerStarted","Data":"92a3d55372681e357aa2ea475e6fe5755b89f057f4bc70b3a9ad92d74679be31"} Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.386295 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.390601 4868 generic.go:334] "Generic (PLEG): container finished" podID="d7e2ea14-2a0a-418c-ab25-4a588d728f5a" containerID="25ef767abfd397a3463059df500f412275ac41704ca813fd6f2f7caf674c01d9" exitCode=0 Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.391114 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q7782" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.393136 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7782" event={"ID":"d7e2ea14-2a0a-418c-ab25-4a588d728f5a","Type":"ContainerDied","Data":"25ef767abfd397a3463059df500f412275ac41704ca813fd6f2f7caf674c01d9"} Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.393237 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7782" event={"ID":"d7e2ea14-2a0a-418c-ab25-4a588d728f5a","Type":"ContainerDied","Data":"7f8ec6b6373e6427938cd7a070a5c99df781f7f7c61ce8076b07594113ec91fa"} Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.393278 4868 scope.go:117] "RemoveContainer" containerID="25ef767abfd397a3463059df500f412275ac41704ca813fd6f2f7caf674c01d9" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.411658 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-b6k5w" podStartSLOduration=3.411634939 podStartE2EDuration="3.411634939s" podCreationTimestamp="2025-12-01 17:42:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:42:05.411132037 +0000 UTC m=+997.782242448" watchObservedRunningTime="2025-12-01 17:42:05.411634939 +0000 UTC m=+997.782745350" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.423287 4868 scope.go:117] "RemoveContainer" containerID="f7ec7b64cd4b149b3a6a9f00756bb77a57db9aef036b0f79ba14c7740b234b56" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.453328 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q7782"] Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.463501 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q7782"] Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.467442 4868 scope.go:117] "RemoveContainer" containerID="ef5be524361fca9d05a4d37d50616c9ed37d89cce7d86bcaa3672d530fb55536" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.498176 4868 scope.go:117] "RemoveContainer" containerID="25ef767abfd397a3463059df500f412275ac41704ca813fd6f2f7caf674c01d9" Dec 01 17:42:05 crc kubenswrapper[4868]: E1201 17:42:05.500298 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25ef767abfd397a3463059df500f412275ac41704ca813fd6f2f7caf674c01d9\": container with ID starting with 25ef767abfd397a3463059df500f412275ac41704ca813fd6f2f7caf674c01d9 not found: ID does not exist" containerID="25ef767abfd397a3463059df500f412275ac41704ca813fd6f2f7caf674c01d9" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.500363 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25ef767abfd397a3463059df500f412275ac41704ca813fd6f2f7caf674c01d9"} err="failed to get container status \"25ef767abfd397a3463059df500f412275ac41704ca813fd6f2f7caf674c01d9\": rpc error: code = NotFound desc = could not find container \"25ef767abfd397a3463059df500f412275ac41704ca813fd6f2f7caf674c01d9\": container with ID starting with 25ef767abfd397a3463059df500f412275ac41704ca813fd6f2f7caf674c01d9 not found: ID does not exist" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.500426 4868 scope.go:117] "RemoveContainer" containerID="f7ec7b64cd4b149b3a6a9f00756bb77a57db9aef036b0f79ba14c7740b234b56" Dec 01 17:42:05 crc kubenswrapper[4868]: E1201 17:42:05.500860 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7ec7b64cd4b149b3a6a9f00756bb77a57db9aef036b0f79ba14c7740b234b56\": container with ID starting with f7ec7b64cd4b149b3a6a9f00756bb77a57db9aef036b0f79ba14c7740b234b56 not found: ID does not exist" containerID="f7ec7b64cd4b149b3a6a9f00756bb77a57db9aef036b0f79ba14c7740b234b56" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.500895 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7ec7b64cd4b149b3a6a9f00756bb77a57db9aef036b0f79ba14c7740b234b56"} err="failed to get container status \"f7ec7b64cd4b149b3a6a9f00756bb77a57db9aef036b0f79ba14c7740b234b56\": rpc error: code = NotFound desc = could not find container \"f7ec7b64cd4b149b3a6a9f00756bb77a57db9aef036b0f79ba14c7740b234b56\": container with ID starting with f7ec7b64cd4b149b3a6a9f00756bb77a57db9aef036b0f79ba14c7740b234b56 not found: ID does not exist" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.500919 4868 scope.go:117] "RemoveContainer" containerID="ef5be524361fca9d05a4d37d50616c9ed37d89cce7d86bcaa3672d530fb55536" Dec 01 17:42:05 crc kubenswrapper[4868]: E1201 17:42:05.501473 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef5be524361fca9d05a4d37d50616c9ed37d89cce7d86bcaa3672d530fb55536\": container with ID starting with ef5be524361fca9d05a4d37d50616c9ed37d89cce7d86bcaa3672d530fb55536 not found: ID does not exist" containerID="ef5be524361fca9d05a4d37d50616c9ed37d89cce7d86bcaa3672d530fb55536" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.501542 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef5be524361fca9d05a4d37d50616c9ed37d89cce7d86bcaa3672d530fb55536"} err="failed to get container status \"ef5be524361fca9d05a4d37d50616c9ed37d89cce7d86bcaa3672d530fb55536\": rpc error: code = NotFound desc = could not find container \"ef5be524361fca9d05a4d37d50616c9ed37d89cce7d86bcaa3672d530fb55536\": container with ID starting with ef5be524361fca9d05a4d37d50616c9ed37d89cce7d86bcaa3672d530fb55536 not found: ID does not exist" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.820402 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.912239 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-b6k5w"] Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.993511 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-tcxlc"] Dec 01 17:42:05 crc kubenswrapper[4868]: E1201 17:42:05.993930 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e2ea14-2a0a-418c-ab25-4a588d728f5a" containerName="extract-utilities" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.993969 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e2ea14-2a0a-418c-ab25-4a588d728f5a" containerName="extract-utilities" Dec 01 17:42:05 crc kubenswrapper[4868]: E1201 17:42:05.994014 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e2ea14-2a0a-418c-ab25-4a588d728f5a" containerName="registry-server" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.994029 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e2ea14-2a0a-418c-ab25-4a588d728f5a" containerName="registry-server" Dec 01 17:42:05 crc kubenswrapper[4868]: E1201 17:42:05.994047 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e2ea14-2a0a-418c-ab25-4a588d728f5a" containerName="extract-content" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.994057 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e2ea14-2a0a-418c-ab25-4a588d728f5a" containerName="extract-content" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.994297 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e2ea14-2a0a-418c-ab25-4a588d728f5a" containerName="registry-server" Dec 01 17:42:05 crc kubenswrapper[4868]: I1201 17:42:05.996156 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.032700 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-tcxlc"] Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.042056 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-tcxlc\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.042119 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-tcxlc\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.042145 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-tcxlc\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.042251 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdv5k\" (UniqueName: \"kubernetes.io/projected/27bdb46a-6c28-4f18-82e7-50d25471969c-kube-api-access-wdv5k\") pod \"dnsmasq-dns-b8fbc5445-tcxlc\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.042274 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-config\") pod \"dnsmasq-dns-b8fbc5445-tcxlc\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.143839 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdv5k\" (UniqueName: \"kubernetes.io/projected/27bdb46a-6c28-4f18-82e7-50d25471969c-kube-api-access-wdv5k\") pod \"dnsmasq-dns-b8fbc5445-tcxlc\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.143896 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-config\") pod \"dnsmasq-dns-b8fbc5445-tcxlc\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.143956 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-tcxlc\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.143991 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-tcxlc\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.144017 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-tcxlc\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.145041 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-tcxlc\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.145079 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-config\") pod \"dnsmasq-dns-b8fbc5445-tcxlc\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.145122 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-tcxlc\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.145501 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-tcxlc\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.198632 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdv5k\" (UniqueName: \"kubernetes.io/projected/27bdb46a-6c28-4f18-82e7-50d25471969c-kube-api-access-wdv5k\") pod \"dnsmasq-dns-b8fbc5445-tcxlc\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.207541 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7e2ea14-2a0a-418c-ab25-4a588d728f5a" path="/var/lib/kubelet/pods/d7e2ea14-2a0a-418c-ab25-4a588d728f5a/volumes" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.208190 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edf1d58a-829d-4acc-be67-7c76bc6ed0b9" path="/var/lib/kubelet/pods/edf1d58a-829d-4acc-be67-7c76bc6ed0b9/volumes" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.249143 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.317285 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:06 crc kubenswrapper[4868]: I1201 17:42:06.819299 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-tcxlc"] Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.151473 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.157455 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.161006 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-br68m" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.161449 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.163183 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.163237 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.174864 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.269343 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z85pw\" (UniqueName: \"kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-kube-api-access-z85pw\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.269405 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-cache\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.269518 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.269552 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-lock\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.269602 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.371782 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.371878 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-lock\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.371938 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.372098 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z85pw\" (UniqueName: \"kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-kube-api-access-z85pw\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.372238 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-cache\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.373303 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-cache\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.374472 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: E1201 17:42:07.375449 4868 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 17:42:07 crc kubenswrapper[4868]: E1201 17:42:07.375510 4868 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 17:42:07 crc kubenswrapper[4868]: E1201 17:42:07.375609 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift podName:4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4 nodeName:}" failed. No retries permitted until 2025-12-01 17:42:07.875577408 +0000 UTC m=+1000.246687849 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift") pod "swift-storage-0" (UID: "4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4") : configmap "swift-ring-files" not found Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.376546 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-lock\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.406956 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z85pw\" (UniqueName: \"kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-kube-api-access-z85pw\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.416179 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.419788 4868 generic.go:334] "Generic (PLEG): container finished" podID="27bdb46a-6c28-4f18-82e7-50d25471969c" containerID="3f079fdf0fdedea58f7a4baae3764b6eae28f2a934a6223b357708c6283f6c73" exitCode=0 Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.420098 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" event={"ID":"27bdb46a-6c28-4f18-82e7-50d25471969c","Type":"ContainerDied","Data":"3f079fdf0fdedea58f7a4baae3764b6eae28f2a934a6223b357708c6283f6c73"} Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.420152 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-b6k5w" podUID="62062537-295a-4cd8-96e6-1b9961bbfa02" containerName="dnsmasq-dns" containerID="cri-o://92a3d55372681e357aa2ea475e6fe5755b89f057f4bc70b3a9ad92d74679be31" gracePeriod=10 Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.420165 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" event={"ID":"27bdb46a-6c28-4f18-82e7-50d25471969c","Type":"ContainerStarted","Data":"bf7c17c03e968e3a6522cada2e58d9eff3791106b6141f6d1ff4434e1f90f029"} Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.709589 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-9c99v"] Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.710781 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.713824 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.714053 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.714263 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.783095 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-bfdvt"] Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.786792 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.797142 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-9c99v"] Dec 01 17:42:07 crc kubenswrapper[4868]: E1201 17:42:07.798047 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-tll86 ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-tll86 ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-9c99v" podUID="30751961-7326-4783-b116-4a5fc7458bf1" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.815015 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bfdvt"] Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.832218 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-9c99v"] Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.889260 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-swiftconf\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.889494 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/71e5e846-0b2a-47d3-b60b-daf23c22af52-ring-data-devices\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.889542 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tll86\" (UniqueName: \"kubernetes.io/projected/30751961-7326-4783-b116-4a5fc7458bf1-kube-api-access-tll86\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.889568 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-combined-ca-bundle\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.889634 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brwxc\" (UniqueName: \"kubernetes.io/projected/71e5e846-0b2a-47d3-b60b-daf23c22af52-kube-api-access-brwxc\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.889699 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71e5e846-0b2a-47d3-b60b-daf23c22af52-scripts\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.889721 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-combined-ca-bundle\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.889738 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-dispersionconf\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.889785 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/71e5e846-0b2a-47d3-b60b-daf23c22af52-etc-swift\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.889822 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30751961-7326-4783-b116-4a5fc7458bf1-scripts\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.889894 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.889972 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-dispersionconf\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:07 crc kubenswrapper[4868]: E1201 17:42:07.890187 4868 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 17:42:07 crc kubenswrapper[4868]: E1201 17:42:07.890207 4868 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 17:42:07 crc kubenswrapper[4868]: E1201 17:42:07.890284 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift podName:4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4 nodeName:}" failed. No retries permitted until 2025-12-01 17:42:08.890265049 +0000 UTC m=+1001.261375530 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift") pod "swift-storage-0" (UID: "4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4") : configmap "swift-ring-files" not found Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.890590 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-swiftconf\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.890703 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30751961-7326-4783-b116-4a5fc7458bf1-ring-data-devices\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.890780 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30751961-7326-4783-b116-4a5fc7458bf1-etc-swift\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.931757 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.992524 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30751961-7326-4783-b116-4a5fc7458bf1-ring-data-devices\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.992610 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30751961-7326-4783-b116-4a5fc7458bf1-etc-swift\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.992674 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-swiftconf\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.992707 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/71e5e846-0b2a-47d3-b60b-daf23c22af52-ring-data-devices\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.992729 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tll86\" (UniqueName: \"kubernetes.io/projected/30751961-7326-4783-b116-4a5fc7458bf1-kube-api-access-tll86\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.992780 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-combined-ca-bundle\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.992821 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brwxc\" (UniqueName: \"kubernetes.io/projected/71e5e846-0b2a-47d3-b60b-daf23c22af52-kube-api-access-brwxc\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.992870 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71e5e846-0b2a-47d3-b60b-daf23c22af52-scripts\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.992896 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-combined-ca-bundle\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.992923 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-dispersionconf\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.992967 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/71e5e846-0b2a-47d3-b60b-daf23c22af52-etc-swift\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.992996 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30751961-7326-4783-b116-4a5fc7458bf1-scripts\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.993083 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-dispersionconf\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.993117 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-swiftconf\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.995213 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30751961-7326-4783-b116-4a5fc7458bf1-ring-data-devices\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:07 crc kubenswrapper[4868]: I1201 17:42:07.995529 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30751961-7326-4783-b116-4a5fc7458bf1-etc-swift\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.000455 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/71e5e846-0b2a-47d3-b60b-daf23c22af52-ring-data-devices\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.001460 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30751961-7326-4783-b116-4a5fc7458bf1-scripts\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.001595 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-swiftconf\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.001810 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/71e5e846-0b2a-47d3-b60b-daf23c22af52-etc-swift\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.002149 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71e5e846-0b2a-47d3-b60b-daf23c22af52-scripts\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.005338 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-swiftconf\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.010413 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-dispersionconf\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.011104 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-combined-ca-bundle\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.011241 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brwxc\" (UniqueName: \"kubernetes.io/projected/71e5e846-0b2a-47d3-b60b-daf23c22af52-kube-api-access-brwxc\") pod \"swift-ring-rebalance-bfdvt\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.012359 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-dispersionconf\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.013000 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-combined-ca-bundle\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.014484 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tll86\" (UniqueName: \"kubernetes.io/projected/30751961-7326-4783-b116-4a5fc7458bf1-kube-api-access-tll86\") pod \"swift-ring-rebalance-9c99v\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.094394 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-config\") pod \"62062537-295a-4cd8-96e6-1b9961bbfa02\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.094478 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-ovsdbserver-nb\") pod \"62062537-295a-4cd8-96e6-1b9961bbfa02\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.094576 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-dns-svc\") pod \"62062537-295a-4cd8-96e6-1b9961bbfa02\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.094706 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-ovsdbserver-sb\") pod \"62062537-295a-4cd8-96e6-1b9961bbfa02\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.094727 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5p8mx\" (UniqueName: \"kubernetes.io/projected/62062537-295a-4cd8-96e6-1b9961bbfa02-kube-api-access-5p8mx\") pod \"62062537-295a-4cd8-96e6-1b9961bbfa02\" (UID: \"62062537-295a-4cd8-96e6-1b9961bbfa02\") " Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.100300 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62062537-295a-4cd8-96e6-1b9961bbfa02-kube-api-access-5p8mx" (OuterVolumeSpecName: "kube-api-access-5p8mx") pod "62062537-295a-4cd8-96e6-1b9961bbfa02" (UID: "62062537-295a-4cd8-96e6-1b9961bbfa02"). InnerVolumeSpecName "kube-api-access-5p8mx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.142694 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "62062537-295a-4cd8-96e6-1b9961bbfa02" (UID: "62062537-295a-4cd8-96e6-1b9961bbfa02"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.143958 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-config" (OuterVolumeSpecName: "config") pod "62062537-295a-4cd8-96e6-1b9961bbfa02" (UID: "62062537-295a-4cd8-96e6-1b9961bbfa02"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.151331 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "62062537-295a-4cd8-96e6-1b9961bbfa02" (UID: "62062537-295a-4cd8-96e6-1b9961bbfa02"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.161212 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "62062537-295a-4cd8-96e6-1b9961bbfa02" (UID: "62062537-295a-4cd8-96e6-1b9961bbfa02"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.196460 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.196494 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5p8mx\" (UniqueName: \"kubernetes.io/projected/62062537-295a-4cd8-96e6-1b9961bbfa02-kube-api-access-5p8mx\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.196509 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.196522 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.196533 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/62062537-295a-4cd8-96e6-1b9961bbfa02-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.223897 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.448073 4868 generic.go:334] "Generic (PLEG): container finished" podID="62062537-295a-4cd8-96e6-1b9961bbfa02" containerID="92a3d55372681e357aa2ea475e6fe5755b89f057f4bc70b3a9ad92d74679be31" exitCode=0 Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.448169 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-b6k5w" event={"ID":"62062537-295a-4cd8-96e6-1b9961bbfa02","Type":"ContainerDied","Data":"92a3d55372681e357aa2ea475e6fe5755b89f057f4bc70b3a9ad92d74679be31"} Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.448690 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-b6k5w" event={"ID":"62062537-295a-4cd8-96e6-1b9961bbfa02","Type":"ContainerDied","Data":"74b0ea2f951729d12600025ccc7396ab6618936851d4f6e74f9ca250027dac7e"} Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.448733 4868 scope.go:117] "RemoveContainer" containerID="92a3d55372681e357aa2ea475e6fe5755b89f057f4bc70b3a9ad92d74679be31" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.448997 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-b6k5w" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.453808 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.455491 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" event={"ID":"27bdb46a-6c28-4f18-82e7-50d25471969c","Type":"ContainerStarted","Data":"d68162bfc8404f519b834a7cb4da209f50f4de3ecef416f60b6be3e708c34a31"} Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.455521 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.474659 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" podStartSLOduration=3.47462488 podStartE2EDuration="3.47462488s" podCreationTimestamp="2025-12-01 17:42:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:42:08.474174259 +0000 UTC m=+1000.845284680" watchObservedRunningTime="2025-12-01 17:42:08.47462488 +0000 UTC m=+1000.845735301" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.492129 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.506436 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-b6k5w"] Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.516243 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-b6k5w"] Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.607786 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30751961-7326-4783-b116-4a5fc7458bf1-scripts\") pod \"30751961-7326-4783-b116-4a5fc7458bf1\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.607986 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-combined-ca-bundle\") pod \"30751961-7326-4783-b116-4a5fc7458bf1\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.608081 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30751961-7326-4783-b116-4a5fc7458bf1-ring-data-devices\") pod \"30751961-7326-4783-b116-4a5fc7458bf1\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.608152 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tll86\" (UniqueName: \"kubernetes.io/projected/30751961-7326-4783-b116-4a5fc7458bf1-kube-api-access-tll86\") pod \"30751961-7326-4783-b116-4a5fc7458bf1\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.608265 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-swiftconf\") pod \"30751961-7326-4783-b116-4a5fc7458bf1\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.608505 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30751961-7326-4783-b116-4a5fc7458bf1-etc-swift\") pod \"30751961-7326-4783-b116-4a5fc7458bf1\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.608534 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30751961-7326-4783-b116-4a5fc7458bf1-scripts" (OuterVolumeSpecName: "scripts") pod "30751961-7326-4783-b116-4a5fc7458bf1" (UID: "30751961-7326-4783-b116-4a5fc7458bf1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.608584 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-dispersionconf\") pod \"30751961-7326-4783-b116-4a5fc7458bf1\" (UID: \"30751961-7326-4783-b116-4a5fc7458bf1\") " Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.608843 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30751961-7326-4783-b116-4a5fc7458bf1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "30751961-7326-4783-b116-4a5fc7458bf1" (UID: "30751961-7326-4783-b116-4a5fc7458bf1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.609445 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30751961-7326-4783-b116-4a5fc7458bf1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "30751961-7326-4783-b116-4a5fc7458bf1" (UID: "30751961-7326-4783-b116-4a5fc7458bf1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.610166 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30751961-7326-4783-b116-4a5fc7458bf1-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.610179 4868 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30751961-7326-4783-b116-4a5fc7458bf1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.610189 4868 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30751961-7326-4783-b116-4a5fc7458bf1-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.615210 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30751961-7326-4783-b116-4a5fc7458bf1-kube-api-access-tll86" (OuterVolumeSpecName: "kube-api-access-tll86") pod "30751961-7326-4783-b116-4a5fc7458bf1" (UID: "30751961-7326-4783-b116-4a5fc7458bf1"). InnerVolumeSpecName "kube-api-access-tll86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.615617 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30751961-7326-4783-b116-4a5fc7458bf1" (UID: "30751961-7326-4783-b116-4a5fc7458bf1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.615995 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "30751961-7326-4783-b116-4a5fc7458bf1" (UID: "30751961-7326-4783-b116-4a5fc7458bf1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.616701 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "30751961-7326-4783-b116-4a5fc7458bf1" (UID: "30751961-7326-4783-b116-4a5fc7458bf1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.711739 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.711772 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tll86\" (UniqueName: \"kubernetes.io/projected/30751961-7326-4783-b116-4a5fc7458bf1-kube-api-access-tll86\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.711785 4868 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.711810 4868 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30751961-7326-4783-b116-4a5fc7458bf1-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.751642 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bfdvt"] Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.915604 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:08 crc kubenswrapper[4868]: E1201 17:42:08.915832 4868 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 17:42:08 crc kubenswrapper[4868]: E1201 17:42:08.915861 4868 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 17:42:08 crc kubenswrapper[4868]: E1201 17:42:08.915936 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift podName:4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4 nodeName:}" failed. No retries permitted until 2025-12-01 17:42:10.915914544 +0000 UTC m=+1003.287024955 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift") pod "swift-storage-0" (UID: "4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4") : configmap "swift-ring-files" not found Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.942483 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-c223-account-create-update-g7wbx"] Dec 01 17:42:08 crc kubenswrapper[4868]: E1201 17:42:08.943041 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62062537-295a-4cd8-96e6-1b9961bbfa02" containerName="dnsmasq-dns" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.943059 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="62062537-295a-4cd8-96e6-1b9961bbfa02" containerName="dnsmasq-dns" Dec 01 17:42:08 crc kubenswrapper[4868]: E1201 17:42:08.943077 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62062537-295a-4cd8-96e6-1b9961bbfa02" containerName="init" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.943082 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="62062537-295a-4cd8-96e6-1b9961bbfa02" containerName="init" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.943308 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="62062537-295a-4cd8-96e6-1b9961bbfa02" containerName="dnsmasq-dns" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.944009 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c223-account-create-update-g7wbx" Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.949533 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c223-account-create-update-g7wbx"] Dec 01 17:42:08 crc kubenswrapper[4868]: I1201 17:42:08.997803 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.041520 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-592xb"] Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.043216 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-592xb" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.057332 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-592xb"] Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.126154 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f1127ca-d13f-46cf-b792-f9854e37586c-operator-scripts\") pod \"glance-c223-account-create-update-g7wbx\" (UID: \"0f1127ca-d13f-46cf-b792-f9854e37586c\") " pod="openstack/glance-c223-account-create-update-g7wbx" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.126203 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9wqx\" (UniqueName: \"kubernetes.io/projected/0f1127ca-d13f-46cf-b792-f9854e37586c-kube-api-access-g9wqx\") pod \"glance-c223-account-create-update-g7wbx\" (UID: \"0f1127ca-d13f-46cf-b792-f9854e37586c\") " pod="openstack/glance-c223-account-create-update-g7wbx" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.228387 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f1127ca-d13f-46cf-b792-f9854e37586c-operator-scripts\") pod \"glance-c223-account-create-update-g7wbx\" (UID: \"0f1127ca-d13f-46cf-b792-f9854e37586c\") " pod="openstack/glance-c223-account-create-update-g7wbx" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.228443 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9wqx\" (UniqueName: \"kubernetes.io/projected/0f1127ca-d13f-46cf-b792-f9854e37586c-kube-api-access-g9wqx\") pod \"glance-c223-account-create-update-g7wbx\" (UID: \"0f1127ca-d13f-46cf-b792-f9854e37586c\") " pod="openstack/glance-c223-account-create-update-g7wbx" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.228487 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpqmn\" (UniqueName: \"kubernetes.io/projected/7f698ec2-ef7a-4614-8087-5b4f3fe94365-kube-api-access-tpqmn\") pod \"glance-db-create-592xb\" (UID: \"7f698ec2-ef7a-4614-8087-5b4f3fe94365\") " pod="openstack/glance-db-create-592xb" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.228535 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f698ec2-ef7a-4614-8087-5b4f3fe94365-operator-scripts\") pod \"glance-db-create-592xb\" (UID: \"7f698ec2-ef7a-4614-8087-5b4f3fe94365\") " pod="openstack/glance-db-create-592xb" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.229180 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f1127ca-d13f-46cf-b792-f9854e37586c-operator-scripts\") pod \"glance-c223-account-create-update-g7wbx\" (UID: \"0f1127ca-d13f-46cf-b792-f9854e37586c\") " pod="openstack/glance-c223-account-create-update-g7wbx" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.247763 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9wqx\" (UniqueName: \"kubernetes.io/projected/0f1127ca-d13f-46cf-b792-f9854e37586c-kube-api-access-g9wqx\") pod \"glance-c223-account-create-update-g7wbx\" (UID: \"0f1127ca-d13f-46cf-b792-f9854e37586c\") " pod="openstack/glance-c223-account-create-update-g7wbx" Dec 01 17:42:09 crc kubenswrapper[4868]: W1201 17:42:09.304400 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71e5e846_0b2a_47d3_b60b_daf23c22af52.slice/crio-e305e68fa1f9436d87c4c6dd2130442d31c0d11285836ebea390f8f6301e37d2 WatchSource:0}: Error finding container e305e68fa1f9436d87c4c6dd2130442d31c0d11285836ebea390f8f6301e37d2: Status 404 returned error can't find the container with id e305e68fa1f9436d87c4c6dd2130442d31c0d11285836ebea390f8f6301e37d2 Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.313377 4868 scope.go:117] "RemoveContainer" containerID="fd35ed0f45997750f2a7b4f15850a6c3b70d2167cddc8f5830248f28cb07140d" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.330443 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpqmn\" (UniqueName: \"kubernetes.io/projected/7f698ec2-ef7a-4614-8087-5b4f3fe94365-kube-api-access-tpqmn\") pod \"glance-db-create-592xb\" (UID: \"7f698ec2-ef7a-4614-8087-5b4f3fe94365\") " pod="openstack/glance-db-create-592xb" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.331103 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f698ec2-ef7a-4614-8087-5b4f3fe94365-operator-scripts\") pod \"glance-db-create-592xb\" (UID: \"7f698ec2-ef7a-4614-8087-5b4f3fe94365\") " pod="openstack/glance-db-create-592xb" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.331616 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c223-account-create-update-g7wbx" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.332060 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f698ec2-ef7a-4614-8087-5b4f3fe94365-operator-scripts\") pod \"glance-db-create-592xb\" (UID: \"7f698ec2-ef7a-4614-8087-5b4f3fe94365\") " pod="openstack/glance-db-create-592xb" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.348594 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpqmn\" (UniqueName: \"kubernetes.io/projected/7f698ec2-ef7a-4614-8087-5b4f3fe94365-kube-api-access-tpqmn\") pod \"glance-db-create-592xb\" (UID: \"7f698ec2-ef7a-4614-8087-5b4f3fe94365\") " pod="openstack/glance-db-create-592xb" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.360406 4868 scope.go:117] "RemoveContainer" containerID="92a3d55372681e357aa2ea475e6fe5755b89f057f4bc70b3a9ad92d74679be31" Dec 01 17:42:09 crc kubenswrapper[4868]: E1201 17:42:09.360786 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92a3d55372681e357aa2ea475e6fe5755b89f057f4bc70b3a9ad92d74679be31\": container with ID starting with 92a3d55372681e357aa2ea475e6fe5755b89f057f4bc70b3a9ad92d74679be31 not found: ID does not exist" containerID="92a3d55372681e357aa2ea475e6fe5755b89f057f4bc70b3a9ad92d74679be31" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.360840 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92a3d55372681e357aa2ea475e6fe5755b89f057f4bc70b3a9ad92d74679be31"} err="failed to get container status \"92a3d55372681e357aa2ea475e6fe5755b89f057f4bc70b3a9ad92d74679be31\": rpc error: code = NotFound desc = could not find container \"92a3d55372681e357aa2ea475e6fe5755b89f057f4bc70b3a9ad92d74679be31\": container with ID starting with 92a3d55372681e357aa2ea475e6fe5755b89f057f4bc70b3a9ad92d74679be31 not found: ID does not exist" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.360867 4868 scope.go:117] "RemoveContainer" containerID="fd35ed0f45997750f2a7b4f15850a6c3b70d2167cddc8f5830248f28cb07140d" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.361203 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-592xb" Dec 01 17:42:09 crc kubenswrapper[4868]: E1201 17:42:09.361269 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd35ed0f45997750f2a7b4f15850a6c3b70d2167cddc8f5830248f28cb07140d\": container with ID starting with fd35ed0f45997750f2a7b4f15850a6c3b70d2167cddc8f5830248f28cb07140d not found: ID does not exist" containerID="fd35ed0f45997750f2a7b4f15850a6c3b70d2167cddc8f5830248f28cb07140d" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.361314 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd35ed0f45997750f2a7b4f15850a6c3b70d2167cddc8f5830248f28cb07140d"} err="failed to get container status \"fd35ed0f45997750f2a7b4f15850a6c3b70d2167cddc8f5830248f28cb07140d\": rpc error: code = NotFound desc = could not find container \"fd35ed0f45997750f2a7b4f15850a6c3b70d2167cddc8f5830248f28cb07140d\": container with ID starting with fd35ed0f45997750f2a7b4f15850a6c3b70d2167cddc8f5830248f28cb07140d not found: ID does not exist" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.479296 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9c99v" Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.481873 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bfdvt" event={"ID":"71e5e846-0b2a-47d3-b60b-daf23c22af52","Type":"ContainerStarted","Data":"e305e68fa1f9436d87c4c6dd2130442d31c0d11285836ebea390f8f6301e37d2"} Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.596899 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-9c99v"] Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.604512 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-9c99v"] Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.806188 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-592xb"] Dec 01 17:42:09 crc kubenswrapper[4868]: W1201 17:42:09.818294 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f698ec2_ef7a_4614_8087_5b4f3fe94365.slice/crio-5eab9443a3fe94721b5a9acf0288dd079e4bddc158e7af83bdb936058e22961c WatchSource:0}: Error finding container 5eab9443a3fe94721b5a9acf0288dd079e4bddc158e7af83bdb936058e22961c: Status 404 returned error can't find the container with id 5eab9443a3fe94721b5a9acf0288dd079e4bddc158e7af83bdb936058e22961c Dec 01 17:42:09 crc kubenswrapper[4868]: I1201 17:42:09.910663 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c223-account-create-update-g7wbx"] Dec 01 17:42:10 crc kubenswrapper[4868]: I1201 17:42:10.184221 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30751961-7326-4783-b116-4a5fc7458bf1" path="/var/lib/kubelet/pods/30751961-7326-4783-b116-4a5fc7458bf1/volumes" Dec 01 17:42:10 crc kubenswrapper[4868]: I1201 17:42:10.185051 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62062537-295a-4cd8-96e6-1b9961bbfa02" path="/var/lib/kubelet/pods/62062537-295a-4cd8-96e6-1b9961bbfa02/volumes" Dec 01 17:42:10 crc kubenswrapper[4868]: I1201 17:42:10.489020 4868 generic.go:334] "Generic (PLEG): container finished" podID="7f698ec2-ef7a-4614-8087-5b4f3fe94365" containerID="f6e962410e6d04cbee7c7f78ac64f6ae6bc7acc8e4033205654f21e6009eaf91" exitCode=0 Dec 01 17:42:10 crc kubenswrapper[4868]: I1201 17:42:10.489097 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-592xb" event={"ID":"7f698ec2-ef7a-4614-8087-5b4f3fe94365","Type":"ContainerDied","Data":"f6e962410e6d04cbee7c7f78ac64f6ae6bc7acc8e4033205654f21e6009eaf91"} Dec 01 17:42:10 crc kubenswrapper[4868]: I1201 17:42:10.489457 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-592xb" event={"ID":"7f698ec2-ef7a-4614-8087-5b4f3fe94365","Type":"ContainerStarted","Data":"5eab9443a3fe94721b5a9acf0288dd079e4bddc158e7af83bdb936058e22961c"} Dec 01 17:42:10 crc kubenswrapper[4868]: I1201 17:42:10.491414 4868 generic.go:334] "Generic (PLEG): container finished" podID="0f1127ca-d13f-46cf-b792-f9854e37586c" containerID="aa77b0e41cb09a8344c67ce86e528c0d2af863c8cf6d415c25ce5334e8e03de7" exitCode=0 Dec 01 17:42:10 crc kubenswrapper[4868]: I1201 17:42:10.491473 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c223-account-create-update-g7wbx" event={"ID":"0f1127ca-d13f-46cf-b792-f9854e37586c","Type":"ContainerDied","Data":"aa77b0e41cb09a8344c67ce86e528c0d2af863c8cf6d415c25ce5334e8e03de7"} Dec 01 17:42:10 crc kubenswrapper[4868]: I1201 17:42:10.491530 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c223-account-create-update-g7wbx" event={"ID":"0f1127ca-d13f-46cf-b792-f9854e37586c","Type":"ContainerStarted","Data":"65a5938d5c8e526513fd7598e7447e7663a51e92ab56c4ef24f47b8ceabaa611"} Dec 01 17:42:10 crc kubenswrapper[4868]: I1201 17:42:10.494197 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d5daf33c-c3fa-4f86-a422-93e7afa1afaa","Type":"ContainerStarted","Data":"dcb65394c030dbca7a01e84645d6f658c59fe19686951b6c41259c6b905ca678"} Dec 01 17:42:10 crc kubenswrapper[4868]: I1201 17:42:10.495278 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 17:42:10 crc kubenswrapper[4868]: I1201 17:42:10.548454 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=23.313315163 podStartE2EDuration="55.548428593s" podCreationTimestamp="2025-12-01 17:41:15 +0000 UTC" firstStartedPulling="2025-12-01 17:41:37.135688353 +0000 UTC m=+969.506798764" lastFinishedPulling="2025-12-01 17:42:09.370801783 +0000 UTC m=+1001.741912194" observedRunningTime="2025-12-01 17:42:10.542485642 +0000 UTC m=+1002.913596053" watchObservedRunningTime="2025-12-01 17:42:10.548428593 +0000 UTC m=+1002.919539004" Dec 01 17:42:10 crc kubenswrapper[4868]: I1201 17:42:10.972782 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:10 crc kubenswrapper[4868]: E1201 17:42:10.973167 4868 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 17:42:10 crc kubenswrapper[4868]: E1201 17:42:10.973192 4868 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 17:42:10 crc kubenswrapper[4868]: E1201 17:42:10.976204 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift podName:4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4 nodeName:}" failed. No retries permitted until 2025-12-01 17:42:14.973271029 +0000 UTC m=+1007.344381440 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift") pod "swift-storage-0" (UID: "4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4") : configmap "swift-ring-files" not found Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.223087 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-6cp5w"] Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.230679 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6cp5w" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.248761 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-6cp5w"] Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.313740 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35b2d43b-53b4-4b3b-abce-5393c44d1db9-operator-scripts\") pod \"keystone-db-create-6cp5w\" (UID: \"35b2d43b-53b4-4b3b-abce-5393c44d1db9\") " pod="openstack/keystone-db-create-6cp5w" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.313776 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2fps\" (UniqueName: \"kubernetes.io/projected/35b2d43b-53b4-4b3b-abce-5393c44d1db9-kube-api-access-q2fps\") pod \"keystone-db-create-6cp5w\" (UID: \"35b2d43b-53b4-4b3b-abce-5393c44d1db9\") " pod="openstack/keystone-db-create-6cp5w" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.338888 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-1f4f-account-create-update-4qr5v"] Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.340486 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1f4f-account-create-update-4qr5v" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.353220 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.356532 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-1f4f-account-create-update-4qr5v"] Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.394119 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-592xb" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.416971 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b47b287b-3cf1-4814-a47a-76bb02275bb8-operator-scripts\") pod \"keystone-1f4f-account-create-update-4qr5v\" (UID: \"b47b287b-3cf1-4814-a47a-76bb02275bb8\") " pod="openstack/keystone-1f4f-account-create-update-4qr5v" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.418929 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35b2d43b-53b4-4b3b-abce-5393c44d1db9-operator-scripts\") pod \"keystone-db-create-6cp5w\" (UID: \"35b2d43b-53b4-4b3b-abce-5393c44d1db9\") " pod="openstack/keystone-db-create-6cp5w" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.419145 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2fps\" (UniqueName: \"kubernetes.io/projected/35b2d43b-53b4-4b3b-abce-5393c44d1db9-kube-api-access-q2fps\") pod \"keystone-db-create-6cp5w\" (UID: \"35b2d43b-53b4-4b3b-abce-5393c44d1db9\") " pod="openstack/keystone-db-create-6cp5w" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.419258 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjwns\" (UniqueName: \"kubernetes.io/projected/b47b287b-3cf1-4814-a47a-76bb02275bb8-kube-api-access-tjwns\") pod \"keystone-1f4f-account-create-update-4qr5v\" (UID: \"b47b287b-3cf1-4814-a47a-76bb02275bb8\") " pod="openstack/keystone-1f4f-account-create-update-4qr5v" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.420981 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35b2d43b-53b4-4b3b-abce-5393c44d1db9-operator-scripts\") pod \"keystone-db-create-6cp5w\" (UID: \"35b2d43b-53b4-4b3b-abce-5393c44d1db9\") " pod="openstack/keystone-db-create-6cp5w" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.445086 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2fps\" (UniqueName: \"kubernetes.io/projected/35b2d43b-53b4-4b3b-abce-5393c44d1db9-kube-api-access-q2fps\") pod \"keystone-db-create-6cp5w\" (UID: \"35b2d43b-53b4-4b3b-abce-5393c44d1db9\") " pod="openstack/keystone-db-create-6cp5w" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.500927 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c223-account-create-update-g7wbx" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.524680 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpqmn\" (UniqueName: \"kubernetes.io/projected/7f698ec2-ef7a-4614-8087-5b4f3fe94365-kube-api-access-tpqmn\") pod \"7f698ec2-ef7a-4614-8087-5b4f3fe94365\" (UID: \"7f698ec2-ef7a-4614-8087-5b4f3fe94365\") " Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.525213 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f698ec2-ef7a-4614-8087-5b4f3fe94365-operator-scripts\") pod \"7f698ec2-ef7a-4614-8087-5b4f3fe94365\" (UID: \"7f698ec2-ef7a-4614-8087-5b4f3fe94365\") " Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.525545 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjwns\" (UniqueName: \"kubernetes.io/projected/b47b287b-3cf1-4814-a47a-76bb02275bb8-kube-api-access-tjwns\") pod \"keystone-1f4f-account-create-update-4qr5v\" (UID: \"b47b287b-3cf1-4814-a47a-76bb02275bb8\") " pod="openstack/keystone-1f4f-account-create-update-4qr5v" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.525744 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b47b287b-3cf1-4814-a47a-76bb02275bb8-operator-scripts\") pod \"keystone-1f4f-account-create-update-4qr5v\" (UID: \"b47b287b-3cf1-4814-a47a-76bb02275bb8\") " pod="openstack/keystone-1f4f-account-create-update-4qr5v" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.526667 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b47b287b-3cf1-4814-a47a-76bb02275bb8-operator-scripts\") pod \"keystone-1f4f-account-create-update-4qr5v\" (UID: \"b47b287b-3cf1-4814-a47a-76bb02275bb8\") " pod="openstack/keystone-1f4f-account-create-update-4qr5v" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.527047 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f698ec2-ef7a-4614-8087-5b4f3fe94365-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7f698ec2-ef7a-4614-8087-5b4f3fe94365" (UID: "7f698ec2-ef7a-4614-8087-5b4f3fe94365"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.529128 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f698ec2-ef7a-4614-8087-5b4f3fe94365-kube-api-access-tpqmn" (OuterVolumeSpecName: "kube-api-access-tpqmn") pod "7f698ec2-ef7a-4614-8087-5b4f3fe94365" (UID: "7f698ec2-ef7a-4614-8087-5b4f3fe94365"). InnerVolumeSpecName "kube-api-access-tpqmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.532901 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-592xb" event={"ID":"7f698ec2-ef7a-4614-8087-5b4f3fe94365","Type":"ContainerDied","Data":"5eab9443a3fe94721b5a9acf0288dd079e4bddc158e7af83bdb936058e22961c"} Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.533024 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5eab9443a3fe94721b5a9acf0288dd079e4bddc158e7af83bdb936058e22961c" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.533498 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-592xb" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.548045 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-8qflm"] Dec 01 17:42:13 crc kubenswrapper[4868]: E1201 17:42:13.549263 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f698ec2-ef7a-4614-8087-5b4f3fe94365" containerName="mariadb-database-create" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.549398 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f698ec2-ef7a-4614-8087-5b4f3fe94365" containerName="mariadb-database-create" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.549467 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjwns\" (UniqueName: \"kubernetes.io/projected/b47b287b-3cf1-4814-a47a-76bb02275bb8-kube-api-access-tjwns\") pod \"keystone-1f4f-account-create-update-4qr5v\" (UID: \"b47b287b-3cf1-4814-a47a-76bb02275bb8\") " pod="openstack/keystone-1f4f-account-create-update-4qr5v" Dec 01 17:42:13 crc kubenswrapper[4868]: E1201 17:42:13.549483 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1127ca-d13f-46cf-b792-f9854e37586c" containerName="mariadb-account-create-update" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.549542 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1127ca-d13f-46cf-b792-f9854e37586c" containerName="mariadb-account-create-update" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.549829 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f1127ca-d13f-46cf-b792-f9854e37586c" containerName="mariadb-account-create-update" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.549857 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f698ec2-ef7a-4614-8087-5b4f3fe94365" containerName="mariadb-database-create" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.550609 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8qflm" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.554495 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c223-account-create-update-g7wbx" event={"ID":"0f1127ca-d13f-46cf-b792-f9854e37586c","Type":"ContainerDied","Data":"65a5938d5c8e526513fd7598e7447e7663a51e92ab56c4ef24f47b8ceabaa611"} Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.554520 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c223-account-create-update-g7wbx" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.554536 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65a5938d5c8e526513fd7598e7447e7663a51e92ab56c4ef24f47b8ceabaa611" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.563515 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-8qflm"] Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.575995 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6cp5w" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.594851 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1f4f-account-create-update-4qr5v" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.627093 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f1127ca-d13f-46cf-b792-f9854e37586c-operator-scripts\") pod \"0f1127ca-d13f-46cf-b792-f9854e37586c\" (UID: \"0f1127ca-d13f-46cf-b792-f9854e37586c\") " Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.627250 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9wqx\" (UniqueName: \"kubernetes.io/projected/0f1127ca-d13f-46cf-b792-f9854e37586c-kube-api-access-g9wqx\") pod \"0f1127ca-d13f-46cf-b792-f9854e37586c\" (UID: \"0f1127ca-d13f-46cf-b792-f9854e37586c\") " Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.628085 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f698ec2-ef7a-4614-8087-5b4f3fe94365-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.628108 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpqmn\" (UniqueName: \"kubernetes.io/projected/7f698ec2-ef7a-4614-8087-5b4f3fe94365-kube-api-access-tpqmn\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.629171 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f1127ca-d13f-46cf-b792-f9854e37586c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0f1127ca-d13f-46cf-b792-f9854e37586c" (UID: "0f1127ca-d13f-46cf-b792-f9854e37586c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.636559 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f1127ca-d13f-46cf-b792-f9854e37586c-kube-api-access-g9wqx" (OuterVolumeSpecName: "kube-api-access-g9wqx") pod "0f1127ca-d13f-46cf-b792-f9854e37586c" (UID: "0f1127ca-d13f-46cf-b792-f9854e37586c"). InnerVolumeSpecName "kube-api-access-g9wqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.645913 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-f1c8-account-create-update-p4sps"] Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.647006 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f1c8-account-create-update-p4sps" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.650180 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.659148 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f1c8-account-create-update-p4sps"] Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.729620 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cba4291a-4880-41fa-81ad-7df9b4287511-operator-scripts\") pod \"placement-db-create-8qflm\" (UID: \"cba4291a-4880-41fa-81ad-7df9b4287511\") " pod="openstack/placement-db-create-8qflm" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.729867 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cab673b6-4d49-47e9-8b7b-54dabbae8e81-operator-scripts\") pod \"placement-f1c8-account-create-update-p4sps\" (UID: \"cab673b6-4d49-47e9-8b7b-54dabbae8e81\") " pod="openstack/placement-f1c8-account-create-update-p4sps" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.730344 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmpnr\" (UniqueName: \"kubernetes.io/projected/cba4291a-4880-41fa-81ad-7df9b4287511-kube-api-access-lmpnr\") pod \"placement-db-create-8qflm\" (UID: \"cba4291a-4880-41fa-81ad-7df9b4287511\") " pod="openstack/placement-db-create-8qflm" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.730411 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p6vd\" (UniqueName: \"kubernetes.io/projected/cab673b6-4d49-47e9-8b7b-54dabbae8e81-kube-api-access-6p6vd\") pod \"placement-f1c8-account-create-update-p4sps\" (UID: \"cab673b6-4d49-47e9-8b7b-54dabbae8e81\") " pod="openstack/placement-f1c8-account-create-update-p4sps" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.730503 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f1127ca-d13f-46cf-b792-f9854e37586c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.730515 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9wqx\" (UniqueName: \"kubernetes.io/projected/0f1127ca-d13f-46cf-b792-f9854e37586c-kube-api-access-g9wqx\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.831777 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cab673b6-4d49-47e9-8b7b-54dabbae8e81-operator-scripts\") pod \"placement-f1c8-account-create-update-p4sps\" (UID: \"cab673b6-4d49-47e9-8b7b-54dabbae8e81\") " pod="openstack/placement-f1c8-account-create-update-p4sps" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.831897 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmpnr\" (UniqueName: \"kubernetes.io/projected/cba4291a-4880-41fa-81ad-7df9b4287511-kube-api-access-lmpnr\") pod \"placement-db-create-8qflm\" (UID: \"cba4291a-4880-41fa-81ad-7df9b4287511\") " pod="openstack/placement-db-create-8qflm" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.831951 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p6vd\" (UniqueName: \"kubernetes.io/projected/cab673b6-4d49-47e9-8b7b-54dabbae8e81-kube-api-access-6p6vd\") pod \"placement-f1c8-account-create-update-p4sps\" (UID: \"cab673b6-4d49-47e9-8b7b-54dabbae8e81\") " pod="openstack/placement-f1c8-account-create-update-p4sps" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.831995 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cba4291a-4880-41fa-81ad-7df9b4287511-operator-scripts\") pod \"placement-db-create-8qflm\" (UID: \"cba4291a-4880-41fa-81ad-7df9b4287511\") " pod="openstack/placement-db-create-8qflm" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.832608 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cab673b6-4d49-47e9-8b7b-54dabbae8e81-operator-scripts\") pod \"placement-f1c8-account-create-update-p4sps\" (UID: \"cab673b6-4d49-47e9-8b7b-54dabbae8e81\") " pod="openstack/placement-f1c8-account-create-update-p4sps" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.832742 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cba4291a-4880-41fa-81ad-7df9b4287511-operator-scripts\") pod \"placement-db-create-8qflm\" (UID: \"cba4291a-4880-41fa-81ad-7df9b4287511\") " pod="openstack/placement-db-create-8qflm" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.850146 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p6vd\" (UniqueName: \"kubernetes.io/projected/cab673b6-4d49-47e9-8b7b-54dabbae8e81-kube-api-access-6p6vd\") pod \"placement-f1c8-account-create-update-p4sps\" (UID: \"cab673b6-4d49-47e9-8b7b-54dabbae8e81\") " pod="openstack/placement-f1c8-account-create-update-p4sps" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.850461 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmpnr\" (UniqueName: \"kubernetes.io/projected/cba4291a-4880-41fa-81ad-7df9b4287511-kube-api-access-lmpnr\") pod \"placement-db-create-8qflm\" (UID: \"cba4291a-4880-41fa-81ad-7df9b4287511\") " pod="openstack/placement-db-create-8qflm" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.922087 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8qflm" Dec 01 17:42:13 crc kubenswrapper[4868]: I1201 17:42:13.966900 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f1c8-account-create-update-p4sps" Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.112882 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-6cp5w"] Dec 01 17:42:14 crc kubenswrapper[4868]: W1201 17:42:14.131277 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35b2d43b_53b4_4b3b_abce_5393c44d1db9.slice/crio-25b86747540309940f707087b36379d357de2ceaa204343f0d63cd9f36223f58 WatchSource:0}: Error finding container 25b86747540309940f707087b36379d357de2ceaa204343f0d63cd9f36223f58: Status 404 returned error can't find the container with id 25b86747540309940f707087b36379d357de2ceaa204343f0d63cd9f36223f58 Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.218562 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-1f4f-account-create-update-4qr5v"] Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.377576 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-8qflm"] Dec 01 17:42:14 crc kubenswrapper[4868]: W1201 17:42:14.384259 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcba4291a_4880_41fa_81ad_7df9b4287511.slice/crio-b3c1adc2e39a0aeaf2ec066c6d9394e113c21cc13a06baf2d9e4b06b9049a4d8 WatchSource:0}: Error finding container b3c1adc2e39a0aeaf2ec066c6d9394e113c21cc13a06baf2d9e4b06b9049a4d8: Status 404 returned error can't find the container with id b3c1adc2e39a0aeaf2ec066c6d9394e113c21cc13a06baf2d9e4b06b9049a4d8 Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.488540 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f1c8-account-create-update-p4sps"] Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.570157 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-6cp5w" event={"ID":"35b2d43b-53b4-4b3b-abce-5393c44d1db9","Type":"ContainerStarted","Data":"6067bb7b9f378fe293a88744787d62b6d33587fd676beec9c748c509edfe2da2"} Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.570236 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-6cp5w" event={"ID":"35b2d43b-53b4-4b3b-abce-5393c44d1db9","Type":"ContainerStarted","Data":"25b86747540309940f707087b36379d357de2ceaa204343f0d63cd9f36223f58"} Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.573264 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1f4f-account-create-update-4qr5v" event={"ID":"b47b287b-3cf1-4814-a47a-76bb02275bb8","Type":"ContainerStarted","Data":"0be7ba16763099d2d5465f8dd9f7c6e29a71a14ca61ceac2525dd920b54b1dd3"} Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.573324 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1f4f-account-create-update-4qr5v" event={"ID":"b47b287b-3cf1-4814-a47a-76bb02275bb8","Type":"ContainerStarted","Data":"ab6d505efb2fa64727fd1fb39187d7265c5a88811da1c67f46f1e1abbc357ffe"} Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.580646 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"faa33685-51f4-46ce-bdc7-ef9a2467f482","Type":"ContainerStarted","Data":"1b5f277dd09d6466063369a51b90654d979e2d28593bba3613046c7d7eabe1f6"} Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.589163 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-6cp5w" podStartSLOduration=1.5891352109999999 podStartE2EDuration="1.589135211s" podCreationTimestamp="2025-12-01 17:42:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:42:14.58909049 +0000 UTC m=+1006.960200901" watchObservedRunningTime="2025-12-01 17:42:14.589135211 +0000 UTC m=+1006.960245622" Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.590681 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"89b376d8-f72b-4c03-a2f1-3fcf6ff20336","Type":"ContainerStarted","Data":"2e1a0d05f5d9d8cae7be7b2cada97501ebb445c909011bede0abce22852e25af"} Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.594463 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f1c8-account-create-update-p4sps" event={"ID":"cab673b6-4d49-47e9-8b7b-54dabbae8e81","Type":"ContainerStarted","Data":"e4ae915a732b6c03c72983bb2cb580a9160b4434fb92f4acaef6000ca5afe87b"} Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.601125 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8qflm" event={"ID":"cba4291a-4880-41fa-81ad-7df9b4287511","Type":"ContainerStarted","Data":"b3c1adc2e39a0aeaf2ec066c6d9394e113c21cc13a06baf2d9e4b06b9049a4d8"} Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.605572 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bfdvt" event={"ID":"71e5e846-0b2a-47d3-b60b-daf23c22af52","Type":"ContainerStarted","Data":"477247bf0de52f409e7ba54b3a149a6faaddf3fafd218af89c49334c5a9187b8"} Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.612884 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-1f4f-account-create-update-4qr5v" podStartSLOduration=1.612860564 podStartE2EDuration="1.612860564s" podCreationTimestamp="2025-12-01 17:42:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:42:14.610720351 +0000 UTC m=+1006.981830772" watchObservedRunningTime="2025-12-01 17:42:14.612860564 +0000 UTC m=+1006.983970975" Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.644160 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=17.681040502 podStartE2EDuration="53.644126349s" podCreationTimestamp="2025-12-01 17:41:21 +0000 UTC" firstStartedPulling="2025-12-01 17:41:37.423633913 +0000 UTC m=+969.794744324" lastFinishedPulling="2025-12-01 17:42:13.38671976 +0000 UTC m=+1005.757830171" observedRunningTime="2025-12-01 17:42:14.636995528 +0000 UTC m=+1007.008105939" watchObservedRunningTime="2025-12-01 17:42:14.644126349 +0000 UTC m=+1007.015236760" Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.666318 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-bfdvt" podStartSLOduration=3.59149687 podStartE2EDuration="7.666294663s" podCreationTimestamp="2025-12-01 17:42:07 +0000 UTC" firstStartedPulling="2025-12-01 17:42:09.313477506 +0000 UTC m=+1001.684587917" lastFinishedPulling="2025-12-01 17:42:13.388275299 +0000 UTC m=+1005.759385710" observedRunningTime="2025-12-01 17:42:14.660294911 +0000 UTC m=+1007.031405322" watchObservedRunningTime="2025-12-01 17:42:14.666294663 +0000 UTC m=+1007.037405074" Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.686339 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=20.529560849 podStartE2EDuration="56.686315033s" podCreationTimestamp="2025-12-01 17:41:18 +0000 UTC" firstStartedPulling="2025-12-01 17:41:37.231379582 +0000 UTC m=+969.602489993" lastFinishedPulling="2025-12-01 17:42:13.388133766 +0000 UTC m=+1005.759244177" observedRunningTime="2025-12-01 17:42:14.680150406 +0000 UTC m=+1007.051260817" watchObservedRunningTime="2025-12-01 17:42:14.686315033 +0000 UTC m=+1007.057425434" Dec 01 17:42:14 crc kubenswrapper[4868]: I1201 17:42:14.834394 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 01 17:42:15 crc kubenswrapper[4868]: I1201 17:42:15.055969 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:15 crc kubenswrapper[4868]: E1201 17:42:15.056264 4868 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 17:42:15 crc kubenswrapper[4868]: E1201 17:42:15.056285 4868 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 17:42:15 crc kubenswrapper[4868]: E1201 17:42:15.056339 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift podName:4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4 nodeName:}" failed. No retries permitted until 2025-12-01 17:42:23.056320773 +0000 UTC m=+1015.427431184 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift") pod "swift-storage-0" (UID: "4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4") : configmap "swift-ring-files" not found Dec 01 17:42:15 crc kubenswrapper[4868]: I1201 17:42:15.620621 4868 generic.go:334] "Generic (PLEG): container finished" podID="b47b287b-3cf1-4814-a47a-76bb02275bb8" containerID="0be7ba16763099d2d5465f8dd9f7c6e29a71a14ca61ceac2525dd920b54b1dd3" exitCode=0 Dec 01 17:42:15 crc kubenswrapper[4868]: I1201 17:42:15.620705 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1f4f-account-create-update-4qr5v" event={"ID":"b47b287b-3cf1-4814-a47a-76bb02275bb8","Type":"ContainerDied","Data":"0be7ba16763099d2d5465f8dd9f7c6e29a71a14ca61ceac2525dd920b54b1dd3"} Dec 01 17:42:15 crc kubenswrapper[4868]: I1201 17:42:15.633446 4868 generic.go:334] "Generic (PLEG): container finished" podID="cba4291a-4880-41fa-81ad-7df9b4287511" containerID="71c96f022e567aaecd6b1e11c5cc6b53c845bd76468079b26cfcdfd944c11d11" exitCode=0 Dec 01 17:42:15 crc kubenswrapper[4868]: I1201 17:42:15.633527 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8qflm" event={"ID":"cba4291a-4880-41fa-81ad-7df9b4287511","Type":"ContainerDied","Data":"71c96f022e567aaecd6b1e11c5cc6b53c845bd76468079b26cfcdfd944c11d11"} Dec 01 17:42:15 crc kubenswrapper[4868]: I1201 17:42:15.646696 4868 generic.go:334] "Generic (PLEG): container finished" podID="cab673b6-4d49-47e9-8b7b-54dabbae8e81" containerID="7cd4dc57fd9bd34ab783eabf2fc54393fd2ee133a1fb481773fc1cdf00c398e2" exitCode=0 Dec 01 17:42:15 crc kubenswrapper[4868]: I1201 17:42:15.646844 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f1c8-account-create-update-p4sps" event={"ID":"cab673b6-4d49-47e9-8b7b-54dabbae8e81","Type":"ContainerDied","Data":"7cd4dc57fd9bd34ab783eabf2fc54393fd2ee133a1fb481773fc1cdf00c398e2"} Dec 01 17:42:15 crc kubenswrapper[4868]: I1201 17:42:15.649234 4868 generic.go:334] "Generic (PLEG): container finished" podID="35b2d43b-53b4-4b3b-abce-5393c44d1db9" containerID="6067bb7b9f378fe293a88744787d62b6d33587fd676beec9c748c509edfe2da2" exitCode=0 Dec 01 17:42:15 crc kubenswrapper[4868]: I1201 17:42:15.650309 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-6cp5w" event={"ID":"35b2d43b-53b4-4b3b-abce-5393c44d1db9","Type":"ContainerDied","Data":"6067bb7b9f378fe293a88744787d62b6d33587fd676beec9c748c509edfe2da2"} Dec 01 17:42:15 crc kubenswrapper[4868]: I1201 17:42:15.778589 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 17:42:16 crc kubenswrapper[4868]: I1201 17:42:16.319095 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:16 crc kubenswrapper[4868]: I1201 17:42:16.398844 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lr5ft"] Dec 01 17:42:16 crc kubenswrapper[4868]: I1201 17:42:16.399169 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" podUID="ba10e6ec-3b28-462a-b625-a79d63eb5f92" containerName="dnsmasq-dns" containerID="cri-o://a70f2cb3ad7c12355fb4be946c123ae616452b53e93b86b1971105d87dc0ebdc" gracePeriod=10 Dec 01 17:42:16 crc kubenswrapper[4868]: I1201 17:42:16.406170 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 01 17:42:16 crc kubenswrapper[4868]: I1201 17:42:16.460821 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 01 17:42:16 crc kubenswrapper[4868]: I1201 17:42:16.666774 4868 generic.go:334] "Generic (PLEG): container finished" podID="ba10e6ec-3b28-462a-b625-a79d63eb5f92" containerID="a70f2cb3ad7c12355fb4be946c123ae616452b53e93b86b1971105d87dc0ebdc" exitCode=0 Dec 01 17:42:16 crc kubenswrapper[4868]: I1201 17:42:16.666935 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" event={"ID":"ba10e6ec-3b28-462a-b625-a79d63eb5f92","Type":"ContainerDied","Data":"a70f2cb3ad7c12355fb4be946c123ae616452b53e93b86b1971105d87dc0ebdc"} Dec 01 17:42:16 crc kubenswrapper[4868]: I1201 17:42:16.667551 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 01 17:42:16 crc kubenswrapper[4868]: I1201 17:42:16.834650 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 01 17:42:16 crc kubenswrapper[4868]: I1201 17:42:16.894725 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" Dec 01 17:42:16 crc kubenswrapper[4868]: I1201 17:42:16.911708 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.002823 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba10e6ec-3b28-462a-b625-a79d63eb5f92-config\") pod \"ba10e6ec-3b28-462a-b625-a79d63eb5f92\" (UID: \"ba10e6ec-3b28-462a-b625-a79d63eb5f92\") " Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.003006 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ghsb\" (UniqueName: \"kubernetes.io/projected/ba10e6ec-3b28-462a-b625-a79d63eb5f92-kube-api-access-8ghsb\") pod \"ba10e6ec-3b28-462a-b625-a79d63eb5f92\" (UID: \"ba10e6ec-3b28-462a-b625-a79d63eb5f92\") " Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.003030 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba10e6ec-3b28-462a-b625-a79d63eb5f92-dns-svc\") pod \"ba10e6ec-3b28-462a-b625-a79d63eb5f92\" (UID: \"ba10e6ec-3b28-462a-b625-a79d63eb5f92\") " Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.012628 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba10e6ec-3b28-462a-b625-a79d63eb5f92-kube-api-access-8ghsb" (OuterVolumeSpecName: "kube-api-access-8ghsb") pod "ba10e6ec-3b28-462a-b625-a79d63eb5f92" (UID: "ba10e6ec-3b28-462a-b625-a79d63eb5f92"). InnerVolumeSpecName "kube-api-access-8ghsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.037986 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1f4f-account-create-update-4qr5v" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.073346 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba10e6ec-3b28-462a-b625-a79d63eb5f92-config" (OuterVolumeSpecName: "config") pod "ba10e6ec-3b28-462a-b625-a79d63eb5f92" (UID: "ba10e6ec-3b28-462a-b625-a79d63eb5f92"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.075779 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba10e6ec-3b28-462a-b625-a79d63eb5f92-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ba10e6ec-3b28-462a-b625-a79d63eb5f92" (UID: "ba10e6ec-3b28-462a-b625-a79d63eb5f92"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.085410 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6cp5w" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.108806 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b47b287b-3cf1-4814-a47a-76bb02275bb8-operator-scripts\") pod \"b47b287b-3cf1-4814-a47a-76bb02275bb8\" (UID: \"b47b287b-3cf1-4814-a47a-76bb02275bb8\") " Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.109053 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjwns\" (UniqueName: \"kubernetes.io/projected/b47b287b-3cf1-4814-a47a-76bb02275bb8-kube-api-access-tjwns\") pod \"b47b287b-3cf1-4814-a47a-76bb02275bb8\" (UID: \"b47b287b-3cf1-4814-a47a-76bb02275bb8\") " Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.109590 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba10e6ec-3b28-462a-b625-a79d63eb5f92-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.109614 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ghsb\" (UniqueName: \"kubernetes.io/projected/ba10e6ec-3b28-462a-b625-a79d63eb5f92-kube-api-access-8ghsb\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.109627 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba10e6ec-3b28-462a-b625-a79d63eb5f92-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.110427 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b47b287b-3cf1-4814-a47a-76bb02275bb8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b47b287b-3cf1-4814-a47a-76bb02275bb8" (UID: "b47b287b-3cf1-4814-a47a-76bb02275bb8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.113332 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b47b287b-3cf1-4814-a47a-76bb02275bb8-kube-api-access-tjwns" (OuterVolumeSpecName: "kube-api-access-tjwns") pod "b47b287b-3cf1-4814-a47a-76bb02275bb8" (UID: "b47b287b-3cf1-4814-a47a-76bb02275bb8"). InnerVolumeSpecName "kube-api-access-tjwns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.194476 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f1c8-account-create-update-p4sps" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.205513 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8qflm" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.213222 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35b2d43b-53b4-4b3b-abce-5393c44d1db9-operator-scripts\") pod \"35b2d43b-53b4-4b3b-abce-5393c44d1db9\" (UID: \"35b2d43b-53b4-4b3b-abce-5393c44d1db9\") " Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.213497 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2fps\" (UniqueName: \"kubernetes.io/projected/35b2d43b-53b4-4b3b-abce-5393c44d1db9-kube-api-access-q2fps\") pod \"35b2d43b-53b4-4b3b-abce-5393c44d1db9\" (UID: \"35b2d43b-53b4-4b3b-abce-5393c44d1db9\") " Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.214592 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b47b287b-3cf1-4814-a47a-76bb02275bb8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.214613 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjwns\" (UniqueName: \"kubernetes.io/projected/b47b287b-3cf1-4814-a47a-76bb02275bb8-kube-api-access-tjwns\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.214713 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35b2d43b-53b4-4b3b-abce-5393c44d1db9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "35b2d43b-53b4-4b3b-abce-5393c44d1db9" (UID: "35b2d43b-53b4-4b3b-abce-5393c44d1db9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.218430 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35b2d43b-53b4-4b3b-abce-5393c44d1db9-kube-api-access-q2fps" (OuterVolumeSpecName: "kube-api-access-q2fps") pod "35b2d43b-53b4-4b3b-abce-5393c44d1db9" (UID: "35b2d43b-53b4-4b3b-abce-5393c44d1db9"). InnerVolumeSpecName "kube-api-access-q2fps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.315296 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cba4291a-4880-41fa-81ad-7df9b4287511-operator-scripts\") pod \"cba4291a-4880-41fa-81ad-7df9b4287511\" (UID: \"cba4291a-4880-41fa-81ad-7df9b4287511\") " Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.315367 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cab673b6-4d49-47e9-8b7b-54dabbae8e81-operator-scripts\") pod \"cab673b6-4d49-47e9-8b7b-54dabbae8e81\" (UID: \"cab673b6-4d49-47e9-8b7b-54dabbae8e81\") " Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.315469 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p6vd\" (UniqueName: \"kubernetes.io/projected/cab673b6-4d49-47e9-8b7b-54dabbae8e81-kube-api-access-6p6vd\") pod \"cab673b6-4d49-47e9-8b7b-54dabbae8e81\" (UID: \"cab673b6-4d49-47e9-8b7b-54dabbae8e81\") " Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.315751 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cba4291a-4880-41fa-81ad-7df9b4287511-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cba4291a-4880-41fa-81ad-7df9b4287511" (UID: "cba4291a-4880-41fa-81ad-7df9b4287511"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.315893 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cab673b6-4d49-47e9-8b7b-54dabbae8e81-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cab673b6-4d49-47e9-8b7b-54dabbae8e81" (UID: "cab673b6-4d49-47e9-8b7b-54dabbae8e81"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.316146 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmpnr\" (UniqueName: \"kubernetes.io/projected/cba4291a-4880-41fa-81ad-7df9b4287511-kube-api-access-lmpnr\") pod \"cba4291a-4880-41fa-81ad-7df9b4287511\" (UID: \"cba4291a-4880-41fa-81ad-7df9b4287511\") " Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.316674 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35b2d43b-53b4-4b3b-abce-5393c44d1db9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.316696 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cba4291a-4880-41fa-81ad-7df9b4287511-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.316708 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cab673b6-4d49-47e9-8b7b-54dabbae8e81-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.316730 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2fps\" (UniqueName: \"kubernetes.io/projected/35b2d43b-53b4-4b3b-abce-5393c44d1db9-kube-api-access-q2fps\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.318319 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cab673b6-4d49-47e9-8b7b-54dabbae8e81-kube-api-access-6p6vd" (OuterVolumeSpecName: "kube-api-access-6p6vd") pod "cab673b6-4d49-47e9-8b7b-54dabbae8e81" (UID: "cab673b6-4d49-47e9-8b7b-54dabbae8e81"). InnerVolumeSpecName "kube-api-access-6p6vd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.319449 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cba4291a-4880-41fa-81ad-7df9b4287511-kube-api-access-lmpnr" (OuterVolumeSpecName: "kube-api-access-lmpnr") pod "cba4291a-4880-41fa-81ad-7df9b4287511" (UID: "cba4291a-4880-41fa-81ad-7df9b4287511"). InnerVolumeSpecName "kube-api-access-lmpnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.417816 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p6vd\" (UniqueName: \"kubernetes.io/projected/cab673b6-4d49-47e9-8b7b-54dabbae8e81-kube-api-access-6p6vd\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.417858 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmpnr\" (UniqueName: \"kubernetes.io/projected/cba4291a-4880-41fa-81ad-7df9b4287511-kube-api-access-lmpnr\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.677394 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f1c8-account-create-update-p4sps" event={"ID":"cab673b6-4d49-47e9-8b7b-54dabbae8e81","Type":"ContainerDied","Data":"e4ae915a732b6c03c72983bb2cb580a9160b4434fb92f4acaef6000ca5afe87b"} Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.677678 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4ae915a732b6c03c72983bb2cb580a9160b4434fb92f4acaef6000ca5afe87b" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.677735 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f1c8-account-create-update-p4sps" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.681415 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8qflm" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.681418 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8qflm" event={"ID":"cba4291a-4880-41fa-81ad-7df9b4287511","Type":"ContainerDied","Data":"b3c1adc2e39a0aeaf2ec066c6d9394e113c21cc13a06baf2d9e4b06b9049a4d8"} Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.681533 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3c1adc2e39a0aeaf2ec066c6d9394e113c21cc13a06baf2d9e4b06b9049a4d8" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.683637 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.683636 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lr5ft" event={"ID":"ba10e6ec-3b28-462a-b625-a79d63eb5f92","Type":"ContainerDied","Data":"e1f514e1b0ed9e46125087bafbf57c9a6ae9b0426bcdbb12242cb45cadc54c35"} Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.683908 4868 scope.go:117] "RemoveContainer" containerID="a70f2cb3ad7c12355fb4be946c123ae616452b53e93b86b1971105d87dc0ebdc" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.685144 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-6cp5w" event={"ID":"35b2d43b-53b4-4b3b-abce-5393c44d1db9","Type":"ContainerDied","Data":"25b86747540309940f707087b36379d357de2ceaa204343f0d63cd9f36223f58"} Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.685181 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25b86747540309940f707087b36379d357de2ceaa204343f0d63cd9f36223f58" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.685206 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6cp5w" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.688252 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1f4f-account-create-update-4qr5v" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.694189 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1f4f-account-create-update-4qr5v" event={"ID":"b47b287b-3cf1-4814-a47a-76bb02275bb8","Type":"ContainerDied","Data":"ab6d505efb2fa64727fd1fb39187d7265c5a88811da1c67f46f1e1abbc357ffe"} Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.694218 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab6d505efb2fa64727fd1fb39187d7265c5a88811da1c67f46f1e1abbc357ffe" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.709339 4868 scope.go:117] "RemoveContainer" containerID="fb74a1d1398a99d7b226741161cc879ac7da92c408cc2bbce5fdba987d7cd177" Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.799718 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lr5ft"] Dec 01 17:42:17 crc kubenswrapper[4868]: I1201 17:42:17.806641 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lr5ft"] Dec 01 17:42:18 crc kubenswrapper[4868]: I1201 17:42:18.183988 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba10e6ec-3b28-462a-b625-a79d63eb5f92" path="/var/lib/kubelet/pods/ba10e6ec-3b28-462a-b625-a79d63eb5f92/volumes" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.245988 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-6z7mt"] Dec 01 17:42:19 crc kubenswrapper[4868]: E1201 17:42:19.246653 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b47b287b-3cf1-4814-a47a-76bb02275bb8" containerName="mariadb-account-create-update" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.246669 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b47b287b-3cf1-4814-a47a-76bb02275bb8" containerName="mariadb-account-create-update" Dec 01 17:42:19 crc kubenswrapper[4868]: E1201 17:42:19.246684 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba10e6ec-3b28-462a-b625-a79d63eb5f92" containerName="dnsmasq-dns" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.246690 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba10e6ec-3b28-462a-b625-a79d63eb5f92" containerName="dnsmasq-dns" Dec 01 17:42:19 crc kubenswrapper[4868]: E1201 17:42:19.246696 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cab673b6-4d49-47e9-8b7b-54dabbae8e81" containerName="mariadb-account-create-update" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.246704 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="cab673b6-4d49-47e9-8b7b-54dabbae8e81" containerName="mariadb-account-create-update" Dec 01 17:42:19 crc kubenswrapper[4868]: E1201 17:42:19.246720 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cba4291a-4880-41fa-81ad-7df9b4287511" containerName="mariadb-database-create" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.246728 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="cba4291a-4880-41fa-81ad-7df9b4287511" containerName="mariadb-database-create" Dec 01 17:42:19 crc kubenswrapper[4868]: E1201 17:42:19.246746 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba10e6ec-3b28-462a-b625-a79d63eb5f92" containerName="init" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.246753 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba10e6ec-3b28-462a-b625-a79d63eb5f92" containerName="init" Dec 01 17:42:19 crc kubenswrapper[4868]: E1201 17:42:19.246776 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35b2d43b-53b4-4b3b-abce-5393c44d1db9" containerName="mariadb-database-create" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.246783 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="35b2d43b-53b4-4b3b-abce-5393c44d1db9" containerName="mariadb-database-create" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.246978 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="cab673b6-4d49-47e9-8b7b-54dabbae8e81" containerName="mariadb-account-create-update" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.246989 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="cba4291a-4880-41fa-81ad-7df9b4287511" containerName="mariadb-database-create" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.247002 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba10e6ec-3b28-462a-b625-a79d63eb5f92" containerName="dnsmasq-dns" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.247062 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="b47b287b-3cf1-4814-a47a-76bb02275bb8" containerName="mariadb-account-create-update" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.247076 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="35b2d43b-53b4-4b3b-abce-5393c44d1db9" containerName="mariadb-database-create" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.247716 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.255031 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5tv8n" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.255274 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.255635 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-combined-ca-bundle\") pod \"glance-db-sync-6z7mt\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.255725 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-config-data\") pod \"glance-db-sync-6z7mt\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.255847 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-db-sync-config-data\") pod \"glance-db-sync-6z7mt\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.255986 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ck7k\" (UniqueName: \"kubernetes.io/projected/a2017518-df87-4ece-a8cf-4805edbf367b-kube-api-access-2ck7k\") pod \"glance-db-sync-6z7mt\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.256278 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-6z7mt"] Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.357852 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ck7k\" (UniqueName: \"kubernetes.io/projected/a2017518-df87-4ece-a8cf-4805edbf367b-kube-api-access-2ck7k\") pod \"glance-db-sync-6z7mt\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.358120 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-combined-ca-bundle\") pod \"glance-db-sync-6z7mt\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.358213 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-config-data\") pod \"glance-db-sync-6z7mt\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.358353 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-db-sync-config-data\") pod \"glance-db-sync-6z7mt\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.368095 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-combined-ca-bundle\") pod \"glance-db-sync-6z7mt\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.368813 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-db-sync-config-data\") pod \"glance-db-sync-6z7mt\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.374504 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-config-data\") pod \"glance-db-sync-6z7mt\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.377471 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ck7k\" (UniqueName: \"kubernetes.io/projected/a2017518-df87-4ece-a8cf-4805edbf367b-kube-api-access-2ck7k\") pod \"glance-db-sync-6z7mt\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.575920 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:19 crc kubenswrapper[4868]: I1201 17:42:19.876221 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 01 17:42:20 crc kubenswrapper[4868]: I1201 17:42:20.116783 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-6z7mt"] Dec 01 17:42:20 crc kubenswrapper[4868]: I1201 17:42:20.717001 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6z7mt" event={"ID":"a2017518-df87-4ece-a8cf-4805edbf367b","Type":"ContainerStarted","Data":"7e52d41b6d7480ea0613a8c70dfdb40b572541c8e6ed510b14030820bceae0df"} Dec 01 17:42:21 crc kubenswrapper[4868]: I1201 17:42:21.728091 4868 generic.go:334] "Generic (PLEG): container finished" podID="71e5e846-0b2a-47d3-b60b-daf23c22af52" containerID="477247bf0de52f409e7ba54b3a149a6faaddf3fafd218af89c49334c5a9187b8" exitCode=0 Dec 01 17:42:21 crc kubenswrapper[4868]: I1201 17:42:21.728197 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bfdvt" event={"ID":"71e5e846-0b2a-47d3-b60b-daf23c22af52","Type":"ContainerDied","Data":"477247bf0de52f409e7ba54b3a149a6faaddf3fafd218af89c49334c5a9187b8"} Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.446990 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.616362 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.626703 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.634623 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.639587 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.639770 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.639848 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.640102 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-s944v" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.716702 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln7vs\" (UniqueName: \"kubernetes.io/projected/56ae3649-0322-453b-9088-6807f59d7d96-kube-api-access-ln7vs\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.716879 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/56ae3649-0322-453b-9088-6807f59d7d96-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.716927 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/56ae3649-0322-453b-9088-6807f59d7d96-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.717085 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/56ae3649-0322-453b-9088-6807f59d7d96-scripts\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.717133 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56ae3649-0322-453b-9088-6807f59d7d96-config\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.717231 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/56ae3649-0322-453b-9088-6807f59d7d96-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.717269 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56ae3649-0322-453b-9088-6807f59d7d96-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.819739 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/56ae3649-0322-453b-9088-6807f59d7d96-scripts\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.819806 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56ae3649-0322-453b-9088-6807f59d7d96-config\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.819874 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/56ae3649-0322-453b-9088-6807f59d7d96-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.819898 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56ae3649-0322-453b-9088-6807f59d7d96-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.819936 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln7vs\" (UniqueName: \"kubernetes.io/projected/56ae3649-0322-453b-9088-6807f59d7d96-kube-api-access-ln7vs\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.819993 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/56ae3649-0322-453b-9088-6807f59d7d96-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.820028 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/56ae3649-0322-453b-9088-6807f59d7d96-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.820642 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/56ae3649-0322-453b-9088-6807f59d7d96-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.822131 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56ae3649-0322-453b-9088-6807f59d7d96-config\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.822804 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/56ae3649-0322-453b-9088-6807f59d7d96-scripts\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.828166 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/56ae3649-0322-453b-9088-6807f59d7d96-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.831885 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56ae3649-0322-453b-9088-6807f59d7d96-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.833105 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/56ae3649-0322-453b-9088-6807f59d7d96-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.854170 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln7vs\" (UniqueName: \"kubernetes.io/projected/56ae3649-0322-453b-9088-6807f59d7d96-kube-api-access-ln7vs\") pod \"ovn-northd-0\" (UID: \"56ae3649-0322-453b-9088-6807f59d7d96\") " pod="openstack/ovn-northd-0" Dec 01 17:42:22 crc kubenswrapper[4868]: I1201 17:42:22.957783 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.073049 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.126498 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.132688 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4-etc-swift\") pod \"swift-storage-0\" (UID: \"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4\") " pod="openstack/swift-storage-0" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.227893 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-dispersionconf\") pod \"71e5e846-0b2a-47d3-b60b-daf23c22af52\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.228030 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-combined-ca-bundle\") pod \"71e5e846-0b2a-47d3-b60b-daf23c22af52\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.228069 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brwxc\" (UniqueName: \"kubernetes.io/projected/71e5e846-0b2a-47d3-b60b-daf23c22af52-kube-api-access-brwxc\") pod \"71e5e846-0b2a-47d3-b60b-daf23c22af52\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.228099 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-swiftconf\") pod \"71e5e846-0b2a-47d3-b60b-daf23c22af52\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.228139 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/71e5e846-0b2a-47d3-b60b-daf23c22af52-ring-data-devices\") pod \"71e5e846-0b2a-47d3-b60b-daf23c22af52\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.228262 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/71e5e846-0b2a-47d3-b60b-daf23c22af52-etc-swift\") pod \"71e5e846-0b2a-47d3-b60b-daf23c22af52\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.228290 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71e5e846-0b2a-47d3-b60b-daf23c22af52-scripts\") pod \"71e5e846-0b2a-47d3-b60b-daf23c22af52\" (UID: \"71e5e846-0b2a-47d3-b60b-daf23c22af52\") " Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.229507 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71e5e846-0b2a-47d3-b60b-daf23c22af52-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "71e5e846-0b2a-47d3-b60b-daf23c22af52" (UID: "71e5e846-0b2a-47d3-b60b-daf23c22af52"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.229755 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71e5e846-0b2a-47d3-b60b-daf23c22af52-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "71e5e846-0b2a-47d3-b60b-daf23c22af52" (UID: "71e5e846-0b2a-47d3-b60b-daf23c22af52"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.233188 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71e5e846-0b2a-47d3-b60b-daf23c22af52-kube-api-access-brwxc" (OuterVolumeSpecName: "kube-api-access-brwxc") pod "71e5e846-0b2a-47d3-b60b-daf23c22af52" (UID: "71e5e846-0b2a-47d3-b60b-daf23c22af52"). InnerVolumeSpecName "kube-api-access-brwxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.235728 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "71e5e846-0b2a-47d3-b60b-daf23c22af52" (UID: "71e5e846-0b2a-47d3-b60b-daf23c22af52"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.252585 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71e5e846-0b2a-47d3-b60b-daf23c22af52-scripts" (OuterVolumeSpecName: "scripts") pod "71e5e846-0b2a-47d3-b60b-daf23c22af52" (UID: "71e5e846-0b2a-47d3-b60b-daf23c22af52"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.258456 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71e5e846-0b2a-47d3-b60b-daf23c22af52" (UID: "71e5e846-0b2a-47d3-b60b-daf23c22af52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.263246 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "71e5e846-0b2a-47d3-b60b-daf23c22af52" (UID: "71e5e846-0b2a-47d3-b60b-daf23c22af52"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.330758 4868 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/71e5e846-0b2a-47d3-b60b-daf23c22af52-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.330849 4868 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/71e5e846-0b2a-47d3-b60b-daf23c22af52-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.330865 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71e5e846-0b2a-47d3-b60b-daf23c22af52-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.330875 4868 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.330910 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.330922 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brwxc\" (UniqueName: \"kubernetes.io/projected/71e5e846-0b2a-47d3-b60b-daf23c22af52-kube-api-access-brwxc\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.330935 4868 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/71e5e846-0b2a-47d3-b60b-daf23c22af52-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.375382 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.579540 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.747880 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bfdvt" event={"ID":"71e5e846-0b2a-47d3-b60b-daf23c22af52","Type":"ContainerDied","Data":"e305e68fa1f9436d87c4c6dd2130442d31c0d11285836ebea390f8f6301e37d2"} Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.747932 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e305e68fa1f9436d87c4c6dd2130442d31c0d11285836ebea390f8f6301e37d2" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.748025 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bfdvt" Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.753391 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"56ae3649-0322-453b-9088-6807f59d7d96","Type":"ContainerStarted","Data":"eb1aaa7ea5ad9c36470b207bdc9c7961950423a5d400c60dbef838de29718042"} Dec 01 17:42:23 crc kubenswrapper[4868]: I1201 17:42:23.994723 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 01 17:42:24 crc kubenswrapper[4868]: W1201 17:42:24.006997 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4997c1c8_2ad2_4b4e_b2ff_6980fcb743e4.slice/crio-119be1421981ab193bc3b4c8ed6bab4cadfe2311c11c309628550c43ed517622 WatchSource:0}: Error finding container 119be1421981ab193bc3b4c8ed6bab4cadfe2311c11c309628550c43ed517622: Status 404 returned error can't find the container with id 119be1421981ab193bc3b4c8ed6bab4cadfe2311c11c309628550c43ed517622 Dec 01 17:42:24 crc kubenswrapper[4868]: I1201 17:42:24.007472 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-z4f4z" podUID="65d7a832-930e-4103-90f6-dbc5de8c1ece" containerName="ovn-controller" probeResult="failure" output=< Dec 01 17:42:24 crc kubenswrapper[4868]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 01 17:42:24 crc kubenswrapper[4868]: > Dec 01 17:42:24 crc kubenswrapper[4868]: I1201 17:42:24.764147 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"119be1421981ab193bc3b4c8ed6bab4cadfe2311c11c309628550c43ed517622"} Dec 01 17:42:25 crc kubenswrapper[4868]: I1201 17:42:25.784469 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"56ae3649-0322-453b-9088-6807f59d7d96","Type":"ContainerStarted","Data":"1d9ec5812b6636a79aaf9b1c7601a837ad50ab9ae8fe62e488c15184b96710bd"} Dec 01 17:42:25 crc kubenswrapper[4868]: I1201 17:42:25.786561 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"03d15e6bfd924a077a64ea0ec1690f3126f7ed51c820350122799e927759f08d"} Dec 01 17:42:25 crc kubenswrapper[4868]: I1201 17:42:25.908481 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:42:25 crc kubenswrapper[4868]: I1201 17:42:25.908543 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:42:26 crc kubenswrapper[4868]: I1201 17:42:26.815763 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"e6ffe2ef5346285c83bfc95b7689580ba64bfb73f05ec413c1f6f479535dc1d1"} Dec 01 17:42:26 crc kubenswrapper[4868]: I1201 17:42:26.816247 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"3d2c6de8d7bd140ba0c19e462be83184a98e74855bdf2c3bc4f1b180f3925cf0"} Dec 01 17:42:26 crc kubenswrapper[4868]: I1201 17:42:26.819469 4868 generic.go:334] "Generic (PLEG): container finished" podID="70aee7bf-d819-4355-a98c-4a106f3e6f16" containerID="571a7ff19cbb3694810ea83e478a19bfa550fa8fb19697462b8c615086834487" exitCode=0 Dec 01 17:42:26 crc kubenswrapper[4868]: I1201 17:42:26.819535 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"70aee7bf-d819-4355-a98c-4a106f3e6f16","Type":"ContainerDied","Data":"571a7ff19cbb3694810ea83e478a19bfa550fa8fb19697462b8c615086834487"} Dec 01 17:42:26 crc kubenswrapper[4868]: I1201 17:42:26.823437 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"56ae3649-0322-453b-9088-6807f59d7d96","Type":"ContainerStarted","Data":"627a956973cd533af0423b1097397504ed66ac7375435c4e94db1a7e5098fc60"} Dec 01 17:42:26 crc kubenswrapper[4868]: I1201 17:42:26.823922 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 01 17:42:26 crc kubenswrapper[4868]: I1201 17:42:26.887201 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.97402138 podStartE2EDuration="4.887182378s" podCreationTimestamp="2025-12-01 17:42:22 +0000 UTC" firstStartedPulling="2025-12-01 17:42:23.586909152 +0000 UTC m=+1015.958019563" lastFinishedPulling="2025-12-01 17:42:25.50007015 +0000 UTC m=+1017.871180561" observedRunningTime="2025-12-01 17:42:26.880543568 +0000 UTC m=+1019.251653979" watchObservedRunningTime="2025-12-01 17:42:26.887182378 +0000 UTC m=+1019.258292789" Dec 01 17:42:27 crc kubenswrapper[4868]: I1201 17:42:27.833300 4868 generic.go:334] "Generic (PLEG): container finished" podID="3ff4f488-06c8-4ceb-848a-4ba68b334752" containerID="bbc47c78ecea10373d709f8016370bad805694e3d2800bf35760bd08eb36dc39" exitCode=0 Dec 01 17:42:27 crc kubenswrapper[4868]: I1201 17:42:27.833397 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3ff4f488-06c8-4ceb-848a-4ba68b334752","Type":"ContainerDied","Data":"bbc47c78ecea10373d709f8016370bad805694e3d2800bf35760bd08eb36dc39"} Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.060395 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-z4f4z" podUID="65d7a832-930e-4103-90f6-dbc5de8c1ece" containerName="ovn-controller" probeResult="failure" output=< Dec 01 17:42:29 crc kubenswrapper[4868]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 01 17:42:29 crc kubenswrapper[4868]: > Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.161559 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.177178 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rzgnl" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.425128 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-z4f4z-config-v6s8q"] Dec 01 17:42:29 crc kubenswrapper[4868]: E1201 17:42:29.426196 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71e5e846-0b2a-47d3-b60b-daf23c22af52" containerName="swift-ring-rebalance" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.426214 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="71e5e846-0b2a-47d3-b60b-daf23c22af52" containerName="swift-ring-rebalance" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.426440 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="71e5e846-0b2a-47d3-b60b-daf23c22af52" containerName="swift-ring-rebalance" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.427360 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.430325 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.433176 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z4f4z-config-v6s8q"] Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.581036 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-run-ovn\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.581165 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-run\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.581206 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6v8m\" (UniqueName: \"kubernetes.io/projected/68082400-d820-484e-a4d5-e8fa7e3d4e97-kube-api-access-x6v8m\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.581844 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68082400-d820-484e-a4d5-e8fa7e3d4e97-scripts\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.582015 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-log-ovn\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.582104 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/68082400-d820-484e-a4d5-e8fa7e3d4e97-additional-scripts\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.685099 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-run-ovn\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.685171 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-run\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.685211 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6v8m\" (UniqueName: \"kubernetes.io/projected/68082400-d820-484e-a4d5-e8fa7e3d4e97-kube-api-access-x6v8m\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.685288 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68082400-d820-484e-a4d5-e8fa7e3d4e97-scripts\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.685335 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-log-ovn\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.685355 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/68082400-d820-484e-a4d5-e8fa7e3d4e97-additional-scripts\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.686856 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/68082400-d820-484e-a4d5-e8fa7e3d4e97-additional-scripts\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.687348 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-run-ovn\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.687400 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-run\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.692267 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-log-ovn\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.692470 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68082400-d820-484e-a4d5-e8fa7e3d4e97-scripts\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.719776 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6v8m\" (UniqueName: \"kubernetes.io/projected/68082400-d820-484e-a4d5-e8fa7e3d4e97-kube-api-access-x6v8m\") pod \"ovn-controller-z4f4z-config-v6s8q\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:29 crc kubenswrapper[4868]: I1201 17:42:29.773242 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:34 crc kubenswrapper[4868]: I1201 17:42:34.001283 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-z4f4z" podUID="65d7a832-930e-4103-90f6-dbc5de8c1ece" containerName="ovn-controller" probeResult="failure" output=< Dec 01 17:42:34 crc kubenswrapper[4868]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 01 17:42:34 crc kubenswrapper[4868]: > Dec 01 17:42:34 crc kubenswrapper[4868]: I1201 17:42:34.139983 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z4f4z-config-v6s8q"] Dec 01 17:42:34 crc kubenswrapper[4868]: I1201 17:42:34.897119 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3ff4f488-06c8-4ceb-848a-4ba68b334752","Type":"ContainerStarted","Data":"a2b5914fe7354791d6773f33ac9378c19a8e5b4afca7990b0b44afa754192933"} Dec 01 17:42:34 crc kubenswrapper[4868]: I1201 17:42:34.898277 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:42:34 crc kubenswrapper[4868]: I1201 17:42:34.901307 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"d87cd7d0547a3574d5e3541ee6e1790af7493c29d14f4f3426b2df4c033832c6"} Dec 01 17:42:34 crc kubenswrapper[4868]: I1201 17:42:34.907038 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6z7mt" event={"ID":"a2017518-df87-4ece-a8cf-4805edbf367b","Type":"ContainerStarted","Data":"3937365a949c713047e17c30856b3a6190c4b98afdcaf20c8a433ab9851f6e3d"} Dec 01 17:42:34 crc kubenswrapper[4868]: I1201 17:42:34.911970 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"70aee7bf-d819-4355-a98c-4a106f3e6f16","Type":"ContainerStarted","Data":"a4829417dd1177e3ef6e9cbc5057cf402a287d114b27dd531e97f35d3beac596"} Dec 01 17:42:34 crc kubenswrapper[4868]: I1201 17:42:34.912313 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 01 17:42:34 crc kubenswrapper[4868]: I1201 17:42:34.914338 4868 generic.go:334] "Generic (PLEG): container finished" podID="68082400-d820-484e-a4d5-e8fa7e3d4e97" containerID="4707587d3ff16d0ec42fd9af5d40088384976db5daa980cb2f8997c715f84231" exitCode=0 Dec 01 17:42:34 crc kubenswrapper[4868]: I1201 17:42:34.914383 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z4f4z-config-v6s8q" event={"ID":"68082400-d820-484e-a4d5-e8fa7e3d4e97","Type":"ContainerDied","Data":"4707587d3ff16d0ec42fd9af5d40088384976db5daa980cb2f8997c715f84231"} Dec 01 17:42:34 crc kubenswrapper[4868]: I1201 17:42:34.914406 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z4f4z-config-v6s8q" event={"ID":"68082400-d820-484e-a4d5-e8fa7e3d4e97","Type":"ContainerStarted","Data":"0918f5b4f87c031c30eab83b01751714954111ee05b5892a77919b1d12dae7d3"} Dec 01 17:42:34 crc kubenswrapper[4868]: I1201 17:42:34.931169 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=69.641363563 podStartE2EDuration="1m25.93114015s" podCreationTimestamp="2025-12-01 17:41:09 +0000 UTC" firstStartedPulling="2025-12-01 17:41:36.621574172 +0000 UTC m=+968.992684583" lastFinishedPulling="2025-12-01 17:41:52.911350759 +0000 UTC m=+985.282461170" observedRunningTime="2025-12-01 17:42:34.92486355 +0000 UTC m=+1027.295973961" watchObservedRunningTime="2025-12-01 17:42:34.93114015 +0000 UTC m=+1027.302250561" Dec 01 17:42:34 crc kubenswrapper[4868]: I1201 17:42:34.956075 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-6z7mt" podStartSLOduration=2.350614835 podStartE2EDuration="15.956050043s" podCreationTimestamp="2025-12-01 17:42:19 +0000 UTC" firstStartedPulling="2025-12-01 17:42:20.127792726 +0000 UTC m=+1012.498903137" lastFinishedPulling="2025-12-01 17:42:33.733227934 +0000 UTC m=+1026.104338345" observedRunningTime="2025-12-01 17:42:34.947505276 +0000 UTC m=+1027.318615707" watchObservedRunningTime="2025-12-01 17:42:34.956050043 +0000 UTC m=+1027.327160454" Dec 01 17:42:34 crc kubenswrapper[4868]: I1201 17:42:34.993996 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=58.967365921 podStartE2EDuration="1m25.993977809s" podCreationTimestamp="2025-12-01 17:41:09 +0000 UTC" firstStartedPulling="2025-12-01 17:41:24.480711011 +0000 UTC m=+956.851821432" lastFinishedPulling="2025-12-01 17:41:51.507322909 +0000 UTC m=+983.878433320" observedRunningTime="2025-12-01 17:42:34.988906229 +0000 UTC m=+1027.360016640" watchObservedRunningTime="2025-12-01 17:42:34.993977809 +0000 UTC m=+1027.365088220" Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.541767 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.658637 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-run-ovn\") pod \"68082400-d820-484e-a4d5-e8fa7e3d4e97\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.659065 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68082400-d820-484e-a4d5-e8fa7e3d4e97-scripts\") pod \"68082400-d820-484e-a4d5-e8fa7e3d4e97\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.659098 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/68082400-d820-484e-a4d5-e8fa7e3d4e97-additional-scripts\") pod \"68082400-d820-484e-a4d5-e8fa7e3d4e97\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.659188 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-run\") pod \"68082400-d820-484e-a4d5-e8fa7e3d4e97\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.659242 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6v8m\" (UniqueName: \"kubernetes.io/projected/68082400-d820-484e-a4d5-e8fa7e3d4e97-kube-api-access-x6v8m\") pod \"68082400-d820-484e-a4d5-e8fa7e3d4e97\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.659297 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-log-ovn\") pod \"68082400-d820-484e-a4d5-e8fa7e3d4e97\" (UID: \"68082400-d820-484e-a4d5-e8fa7e3d4e97\") " Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.659834 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "68082400-d820-484e-a4d5-e8fa7e3d4e97" (UID: "68082400-d820-484e-a4d5-e8fa7e3d4e97"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.659884 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "68082400-d820-484e-a4d5-e8fa7e3d4e97" (UID: "68082400-d820-484e-a4d5-e8fa7e3d4e97"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.660916 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-run" (OuterVolumeSpecName: "var-run") pod "68082400-d820-484e-a4d5-e8fa7e3d4e97" (UID: "68082400-d820-484e-a4d5-e8fa7e3d4e97"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.661371 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68082400-d820-484e-a4d5-e8fa7e3d4e97-scripts" (OuterVolumeSpecName: "scripts") pod "68082400-d820-484e-a4d5-e8fa7e3d4e97" (UID: "68082400-d820-484e-a4d5-e8fa7e3d4e97"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.661474 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68082400-d820-484e-a4d5-e8fa7e3d4e97-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "68082400-d820-484e-a4d5-e8fa7e3d4e97" (UID: "68082400-d820-484e-a4d5-e8fa7e3d4e97"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.693157 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68082400-d820-484e-a4d5-e8fa7e3d4e97-kube-api-access-x6v8m" (OuterVolumeSpecName: "kube-api-access-x6v8m") pod "68082400-d820-484e-a4d5-e8fa7e3d4e97" (UID: "68082400-d820-484e-a4d5-e8fa7e3d4e97"). InnerVolumeSpecName "kube-api-access-x6v8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.761128 4868 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-run\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.761170 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6v8m\" (UniqueName: \"kubernetes.io/projected/68082400-d820-484e-a4d5-e8fa7e3d4e97-kube-api-access-x6v8m\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.761182 4868 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.761192 4868 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/68082400-d820-484e-a4d5-e8fa7e3d4e97-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.761200 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68082400-d820-484e-a4d5-e8fa7e3d4e97-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.761209 4868 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/68082400-d820-484e-a4d5-e8fa7e3d4e97-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.943711 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"764892e081b202125d4687da882883b6fc129540725911a7efc127ccc3c5f081"} Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.946292 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z4f4z-config-v6s8q" Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.946304 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z4f4z-config-v6s8q" event={"ID":"68082400-d820-484e-a4d5-e8fa7e3d4e97","Type":"ContainerDied","Data":"0918f5b4f87c031c30eab83b01751714954111ee05b5892a77919b1d12dae7d3"} Dec 01 17:42:37 crc kubenswrapper[4868]: I1201 17:42:37.946713 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0918f5b4f87c031c30eab83b01751714954111ee05b5892a77919b1d12dae7d3" Dec 01 17:42:38 crc kubenswrapper[4868]: I1201 17:42:38.031983 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 01 17:42:38 crc kubenswrapper[4868]: I1201 17:42:38.671571 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-z4f4z-config-v6s8q"] Dec 01 17:42:38 crc kubenswrapper[4868]: I1201 17:42:38.684116 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-z4f4z-config-v6s8q"] Dec 01 17:42:38 crc kubenswrapper[4868]: I1201 17:42:38.956217 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"5c8dec1148d70009fa3abc0d23723d314d8eea40d1eb0df8ae25a0737712a6a7"} Dec 01 17:42:38 crc kubenswrapper[4868]: I1201 17:42:38.956282 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"c5fe2e73ac08ff98e889b2d631f526fd962b11600b395e424b546d14095c4b22"} Dec 01 17:42:38 crc kubenswrapper[4868]: I1201 17:42:38.956294 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"8494fdb23e6191c5f83c9ab82cffd3041e394ad644cc27ca73855137dd8597a7"} Dec 01 17:42:38 crc kubenswrapper[4868]: I1201 17:42:38.998132 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-z4f4z" Dec 01 17:42:40 crc kubenswrapper[4868]: I1201 17:42:40.182674 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68082400-d820-484e-a4d5-e8fa7e3d4e97" path="/var/lib/kubelet/pods/68082400-d820-484e-a4d5-e8fa7e3d4e97/volumes" Dec 01 17:42:40 crc kubenswrapper[4868]: I1201 17:42:40.979347 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"22d09cf106aa2a021b3cb1445fb2affecd58efdfd8f340d2ac019e86c7c56d7c"} Dec 01 17:42:40 crc kubenswrapper[4868]: I1201 17:42:40.979691 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"6ab6daecffd673ccd595610c47dcc49ccb94a3d4dc257c8808985d203278461b"} Dec 01 17:42:40 crc kubenswrapper[4868]: I1201 17:42:40.979708 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"31633b876f3660aa2d1f2e81f597a35aecc1d4d4e86737d86708162964fc0218"} Dec 01 17:42:40 crc kubenswrapper[4868]: I1201 17:42:40.979723 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"0dc2abbd7e9901957ecf2c3a562746f18053a8f20d3f8039e7294084bb13f998"} Dec 01 17:42:41 crc kubenswrapper[4868]: I1201 17:42:41.994808 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"4ad7c9dd88d95c5f45208fdc9a9877cdcd686a12b3aa2bb366f8a875ade9934d"} Dec 01 17:42:41 crc kubenswrapper[4868]: I1201 17:42:41.995177 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"2edcca252dbd9830dd6ed97f12c7a985f24b59dad32dfd578e0cf7b39d6465e2"} Dec 01 17:42:41 crc kubenswrapper[4868]: I1201 17:42:41.995196 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4","Type":"ContainerStarted","Data":"f2afb36e67ae825ac43b02983a54d3ba78e16b1b05446cb3ca8a31c81ce5c0b5"} Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.034542 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.017953841 podStartE2EDuration="36.034518491s" podCreationTimestamp="2025-12-01 17:42:06 +0000 UTC" firstStartedPulling="2025-12-01 17:42:24.009537951 +0000 UTC m=+1016.380648352" lastFinishedPulling="2025-12-01 17:42:40.026102591 +0000 UTC m=+1032.397213002" observedRunningTime="2025-12-01 17:42:42.030776466 +0000 UTC m=+1034.401886897" watchObservedRunningTime="2025-12-01 17:42:42.034518491 +0000 UTC m=+1034.405628902" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.299032 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-g2gmb"] Dec 01 17:42:42 crc kubenswrapper[4868]: E1201 17:42:42.299517 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68082400-d820-484e-a4d5-e8fa7e3d4e97" containerName="ovn-config" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.299539 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="68082400-d820-484e-a4d5-e8fa7e3d4e97" containerName="ovn-config" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.299755 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="68082400-d820-484e-a4d5-e8fa7e3d4e97" containerName="ovn-config" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.300958 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.304368 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.332757 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-g2gmb"] Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.451619 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.451721 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.451756 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.451788 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.451831 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk6hd\" (UniqueName: \"kubernetes.io/projected/903a0592-08cd-4d36-b634-87cadfb5c944-kube-api-access-nk6hd\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.451876 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-config\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.553334 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.553393 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.553423 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.553456 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk6hd\" (UniqueName: \"kubernetes.io/projected/903a0592-08cd-4d36-b634-87cadfb5c944-kube-api-access-nk6hd\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.553499 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-config\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.553545 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.554449 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.554994 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.555494 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.556383 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.556969 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-config\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.578200 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk6hd\" (UniqueName: \"kubernetes.io/projected/903a0592-08cd-4d36-b634-87cadfb5c944-kube-api-access-nk6hd\") pod \"dnsmasq-dns-5c79d794d7-g2gmb\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:42 crc kubenswrapper[4868]: I1201 17:42:42.617467 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:43 crc kubenswrapper[4868]: I1201 17:42:43.147908 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-g2gmb"] Dec 01 17:42:44 crc kubenswrapper[4868]: I1201 17:42:44.010123 4868 generic.go:334] "Generic (PLEG): container finished" podID="903a0592-08cd-4d36-b634-87cadfb5c944" containerID="e24b0d66899520aaa2e5e5ac35fa6901a0f5185045551bbc5dfe3d0b9e24a37c" exitCode=0 Dec 01 17:42:44 crc kubenswrapper[4868]: I1201 17:42:44.010605 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" event={"ID":"903a0592-08cd-4d36-b634-87cadfb5c944","Type":"ContainerDied","Data":"e24b0d66899520aaa2e5e5ac35fa6901a0f5185045551bbc5dfe3d0b9e24a37c"} Dec 01 17:42:44 crc kubenswrapper[4868]: I1201 17:42:44.010638 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" event={"ID":"903a0592-08cd-4d36-b634-87cadfb5c944","Type":"ContainerStarted","Data":"e50990a355641ed7d0998a27e4e56b359a7cacf119b1c122e22791b658dc2081"} Dec 01 17:42:44 crc kubenswrapper[4868]: I1201 17:42:44.013867 4868 generic.go:334] "Generic (PLEG): container finished" podID="a2017518-df87-4ece-a8cf-4805edbf367b" containerID="3937365a949c713047e17c30856b3a6190c4b98afdcaf20c8a433ab9851f6e3d" exitCode=0 Dec 01 17:42:44 crc kubenswrapper[4868]: I1201 17:42:44.013901 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6z7mt" event={"ID":"a2017518-df87-4ece-a8cf-4805edbf367b","Type":"ContainerDied","Data":"3937365a949c713047e17c30856b3a6190c4b98afdcaf20c8a433ab9851f6e3d"} Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.026241 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" event={"ID":"903a0592-08cd-4d36-b634-87cadfb5c944","Type":"ContainerStarted","Data":"61bd679407463d4b8c3b72ca276962b37c5f35aa0fe7f926d83f9acac92bc0a6"} Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.026586 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.057741 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" podStartSLOduration=3.05771147 podStartE2EDuration="3.05771147s" podCreationTimestamp="2025-12-01 17:42:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:42:45.046192327 +0000 UTC m=+1037.417302738" watchObservedRunningTime="2025-12-01 17:42:45.05771147 +0000 UTC m=+1037.428821891" Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.550852 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.716641 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ck7k\" (UniqueName: \"kubernetes.io/projected/a2017518-df87-4ece-a8cf-4805edbf367b-kube-api-access-2ck7k\") pod \"a2017518-df87-4ece-a8cf-4805edbf367b\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.716869 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-combined-ca-bundle\") pod \"a2017518-df87-4ece-a8cf-4805edbf367b\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.716902 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-config-data\") pod \"a2017518-df87-4ece-a8cf-4805edbf367b\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.716958 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-db-sync-config-data\") pod \"a2017518-df87-4ece-a8cf-4805edbf367b\" (UID: \"a2017518-df87-4ece-a8cf-4805edbf367b\") " Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.724156 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2017518-df87-4ece-a8cf-4805edbf367b-kube-api-access-2ck7k" (OuterVolumeSpecName: "kube-api-access-2ck7k") pod "a2017518-df87-4ece-a8cf-4805edbf367b" (UID: "a2017518-df87-4ece-a8cf-4805edbf367b"). InnerVolumeSpecName "kube-api-access-2ck7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.725169 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a2017518-df87-4ece-a8cf-4805edbf367b" (UID: "a2017518-df87-4ece-a8cf-4805edbf367b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.744250 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a2017518-df87-4ece-a8cf-4805edbf367b" (UID: "a2017518-df87-4ece-a8cf-4805edbf367b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.772853 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-config-data" (OuterVolumeSpecName: "config-data") pod "a2017518-df87-4ece-a8cf-4805edbf367b" (UID: "a2017518-df87-4ece-a8cf-4805edbf367b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.820166 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ck7k\" (UniqueName: \"kubernetes.io/projected/a2017518-df87-4ece-a8cf-4805edbf367b-kube-api-access-2ck7k\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.820236 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.820259 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:45 crc kubenswrapper[4868]: I1201 17:42:45.820279 4868 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a2017518-df87-4ece-a8cf-4805edbf367b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.035201 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6z7mt" event={"ID":"a2017518-df87-4ece-a8cf-4805edbf367b","Type":"ContainerDied","Data":"7e52d41b6d7480ea0613a8c70dfdb40b572541c8e6ed510b14030820bceae0df"} Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.035263 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e52d41b6d7480ea0613a8c70dfdb40b572541c8e6ed510b14030820bceae0df" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.035267 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6z7mt" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.524289 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-g2gmb"] Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.560306 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-ksz2p"] Dec 01 17:42:46 crc kubenswrapper[4868]: E1201 17:42:46.560745 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2017518-df87-4ece-a8cf-4805edbf367b" containerName="glance-db-sync" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.560766 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2017518-df87-4ece-a8cf-4805edbf367b" containerName="glance-db-sync" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.560975 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2017518-df87-4ece-a8cf-4805edbf367b" containerName="glance-db-sync" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.564143 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.578314 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-ksz2p"] Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.736309 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.736709 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.737046 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-config\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.737090 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdlgm\" (UniqueName: \"kubernetes.io/projected/aefbe0d0-2918-4030-8ceb-7e686973efdb-kube-api-access-gdlgm\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.737225 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.737264 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.839249 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.839345 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.839388 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-config\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.839408 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdlgm\" (UniqueName: \"kubernetes.io/projected/aefbe0d0-2918-4030-8ceb-7e686973efdb-kube-api-access-gdlgm\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.839570 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.839594 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.840720 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.841329 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.842001 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-config\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.842278 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.842909 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.864375 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdlgm\" (UniqueName: \"kubernetes.io/projected/aefbe0d0-2918-4030-8ceb-7e686973efdb-kube-api-access-gdlgm\") pod \"dnsmasq-dns-5f59b8f679-ksz2p\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:46 crc kubenswrapper[4868]: I1201 17:42:46.886040 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.050014 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" podUID="903a0592-08cd-4d36-b634-87cadfb5c944" containerName="dnsmasq-dns" containerID="cri-o://61bd679407463d4b8c3b72ca276962b37c5f35aa0fe7f926d83f9acac92bc0a6" gracePeriod=10 Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.371433 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-ksz2p"] Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.480775 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.654757 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-ovsdbserver-sb\") pod \"903a0592-08cd-4d36-b634-87cadfb5c944\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.654935 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-ovsdbserver-nb\") pod \"903a0592-08cd-4d36-b634-87cadfb5c944\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.655011 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nk6hd\" (UniqueName: \"kubernetes.io/projected/903a0592-08cd-4d36-b634-87cadfb5c944-kube-api-access-nk6hd\") pod \"903a0592-08cd-4d36-b634-87cadfb5c944\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.655044 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-config\") pod \"903a0592-08cd-4d36-b634-87cadfb5c944\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.655088 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-dns-svc\") pod \"903a0592-08cd-4d36-b634-87cadfb5c944\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.655157 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-dns-swift-storage-0\") pod \"903a0592-08cd-4d36-b634-87cadfb5c944\" (UID: \"903a0592-08cd-4d36-b634-87cadfb5c944\") " Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.660349 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/903a0592-08cd-4d36-b634-87cadfb5c944-kube-api-access-nk6hd" (OuterVolumeSpecName: "kube-api-access-nk6hd") pod "903a0592-08cd-4d36-b634-87cadfb5c944" (UID: "903a0592-08cd-4d36-b634-87cadfb5c944"). InnerVolumeSpecName "kube-api-access-nk6hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.702806 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "903a0592-08cd-4d36-b634-87cadfb5c944" (UID: "903a0592-08cd-4d36-b634-87cadfb5c944"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.704036 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "903a0592-08cd-4d36-b634-87cadfb5c944" (UID: "903a0592-08cd-4d36-b634-87cadfb5c944"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.710589 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "903a0592-08cd-4d36-b634-87cadfb5c944" (UID: "903a0592-08cd-4d36-b634-87cadfb5c944"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.712485 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "903a0592-08cd-4d36-b634-87cadfb5c944" (UID: "903a0592-08cd-4d36-b634-87cadfb5c944"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.715384 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-config" (OuterVolumeSpecName: "config") pod "903a0592-08cd-4d36-b634-87cadfb5c944" (UID: "903a0592-08cd-4d36-b634-87cadfb5c944"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.757923 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.757998 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.760880 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.760906 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nk6hd\" (UniqueName: \"kubernetes.io/projected/903a0592-08cd-4d36-b634-87cadfb5c944-kube-api-access-nk6hd\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.760923 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:47 crc kubenswrapper[4868]: I1201 17:42:47.760968 4868 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/903a0592-08cd-4d36-b634-87cadfb5c944-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:48 crc kubenswrapper[4868]: I1201 17:42:48.064537 4868 generic.go:334] "Generic (PLEG): container finished" podID="903a0592-08cd-4d36-b634-87cadfb5c944" containerID="61bd679407463d4b8c3b72ca276962b37c5f35aa0fe7f926d83f9acac92bc0a6" exitCode=0 Dec 01 17:42:48 crc kubenswrapper[4868]: I1201 17:42:48.064607 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:42:48 crc kubenswrapper[4868]: I1201 17:42:48.064651 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" event={"ID":"903a0592-08cd-4d36-b634-87cadfb5c944","Type":"ContainerDied","Data":"61bd679407463d4b8c3b72ca276962b37c5f35aa0fe7f926d83f9acac92bc0a6"} Dec 01 17:42:48 crc kubenswrapper[4868]: I1201 17:42:48.064703 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" event={"ID":"903a0592-08cd-4d36-b634-87cadfb5c944","Type":"ContainerDied","Data":"e50990a355641ed7d0998a27e4e56b359a7cacf119b1c122e22791b658dc2081"} Dec 01 17:42:48 crc kubenswrapper[4868]: I1201 17:42:48.064727 4868 scope.go:117] "RemoveContainer" containerID="61bd679407463d4b8c3b72ca276962b37c5f35aa0fe7f926d83f9acac92bc0a6" Dec 01 17:42:48 crc kubenswrapper[4868]: I1201 17:42:48.068326 4868 generic.go:334] "Generic (PLEG): container finished" podID="aefbe0d0-2918-4030-8ceb-7e686973efdb" containerID="79dceeb275033ec3ee9c4e21e84ff0bf283589f4a030667c1cfb7c541c91e301" exitCode=0 Dec 01 17:42:48 crc kubenswrapper[4868]: I1201 17:42:48.068372 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" event={"ID":"aefbe0d0-2918-4030-8ceb-7e686973efdb","Type":"ContainerDied","Data":"79dceeb275033ec3ee9c4e21e84ff0bf283589f4a030667c1cfb7c541c91e301"} Dec 01 17:42:48 crc kubenswrapper[4868]: I1201 17:42:48.068405 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" event={"ID":"aefbe0d0-2918-4030-8ceb-7e686973efdb","Type":"ContainerStarted","Data":"bfa4a88cff9f4c57b326525b5b7a9fe41c02e031260147985bff692754ba5c67"} Dec 01 17:42:48 crc kubenswrapper[4868]: I1201 17:42:48.103235 4868 scope.go:117] "RemoveContainer" containerID="e24b0d66899520aaa2e5e5ac35fa6901a0f5185045551bbc5dfe3d0b9e24a37c" Dec 01 17:42:48 crc kubenswrapper[4868]: I1201 17:42:48.223890 4868 scope.go:117] "RemoveContainer" containerID="61bd679407463d4b8c3b72ca276962b37c5f35aa0fe7f926d83f9acac92bc0a6" Dec 01 17:42:48 crc kubenswrapper[4868]: E1201 17:42:48.224394 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61bd679407463d4b8c3b72ca276962b37c5f35aa0fe7f926d83f9acac92bc0a6\": container with ID starting with 61bd679407463d4b8c3b72ca276962b37c5f35aa0fe7f926d83f9acac92bc0a6 not found: ID does not exist" containerID="61bd679407463d4b8c3b72ca276962b37c5f35aa0fe7f926d83f9acac92bc0a6" Dec 01 17:42:48 crc kubenswrapper[4868]: I1201 17:42:48.224426 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61bd679407463d4b8c3b72ca276962b37c5f35aa0fe7f926d83f9acac92bc0a6"} err="failed to get container status \"61bd679407463d4b8c3b72ca276962b37c5f35aa0fe7f926d83f9acac92bc0a6\": rpc error: code = NotFound desc = could not find container \"61bd679407463d4b8c3b72ca276962b37c5f35aa0fe7f926d83f9acac92bc0a6\": container with ID starting with 61bd679407463d4b8c3b72ca276962b37c5f35aa0fe7f926d83f9acac92bc0a6 not found: ID does not exist" Dec 01 17:42:48 crc kubenswrapper[4868]: I1201 17:42:48.224450 4868 scope.go:117] "RemoveContainer" containerID="e24b0d66899520aaa2e5e5ac35fa6901a0f5185045551bbc5dfe3d0b9e24a37c" Dec 01 17:42:48 crc kubenswrapper[4868]: E1201 17:42:48.224638 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e24b0d66899520aaa2e5e5ac35fa6901a0f5185045551bbc5dfe3d0b9e24a37c\": container with ID starting with e24b0d66899520aaa2e5e5ac35fa6901a0f5185045551bbc5dfe3d0b9e24a37c not found: ID does not exist" containerID="e24b0d66899520aaa2e5e5ac35fa6901a0f5185045551bbc5dfe3d0b9e24a37c" Dec 01 17:42:48 crc kubenswrapper[4868]: I1201 17:42:48.224659 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e24b0d66899520aaa2e5e5ac35fa6901a0f5185045551bbc5dfe3d0b9e24a37c"} err="failed to get container status \"e24b0d66899520aaa2e5e5ac35fa6901a0f5185045551bbc5dfe3d0b9e24a37c\": rpc error: code = NotFound desc = could not find container \"e24b0d66899520aaa2e5e5ac35fa6901a0f5185045551bbc5dfe3d0b9e24a37c\": container with ID starting with e24b0d66899520aaa2e5e5ac35fa6901a0f5185045551bbc5dfe3d0b9e24a37c not found: ID does not exist" Dec 01 17:42:49 crc kubenswrapper[4868]: I1201 17:42:49.089611 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" event={"ID":"aefbe0d0-2918-4030-8ceb-7e686973efdb","Type":"ContainerStarted","Data":"5abd2c337bc714a1a7939c3338e475adc671e7603ac1e9485f7bfd15b6f066ed"} Dec 01 17:42:49 crc kubenswrapper[4868]: I1201 17:42:49.091393 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:49 crc kubenswrapper[4868]: I1201 17:42:49.128719 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" podStartSLOduration=3.128701537 podStartE2EDuration="3.128701537s" podCreationTimestamp="2025-12-01 17:42:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:42:49.125446074 +0000 UTC m=+1041.496556485" watchObservedRunningTime="2025-12-01 17:42:49.128701537 +0000 UTC m=+1041.499811948" Dec 01 17:42:50 crc kubenswrapper[4868]: I1201 17:42:50.630241 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 01 17:42:50 crc kubenswrapper[4868]: I1201 17:42:50.979451 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.005095 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-6b6d-account-create-update-jnz42"] Dec 01 17:42:51 crc kubenswrapper[4868]: E1201 17:42:51.005590 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="903a0592-08cd-4d36-b634-87cadfb5c944" containerName="init" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.005628 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="903a0592-08cd-4d36-b634-87cadfb5c944" containerName="init" Dec 01 17:42:51 crc kubenswrapper[4868]: E1201 17:42:51.005657 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="903a0592-08cd-4d36-b634-87cadfb5c944" containerName="dnsmasq-dns" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.005667 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="903a0592-08cd-4d36-b634-87cadfb5c944" containerName="dnsmasq-dns" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.005893 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="903a0592-08cd-4d36-b634-87cadfb5c944" containerName="dnsmasq-dns" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.006609 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-9zxn4"] Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.007618 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9zxn4" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.007933 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6b6d-account-create-update-jnz42" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.011056 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.016346 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-9zxn4"] Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.023353 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6b6d-account-create-update-jnz42"] Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.099631 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-hmdjz"] Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.102183 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-hmdjz" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.118711 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-hmdjz"] Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.169383 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d9c1c1c-859d-4b2a-9e6b-1a3d863df282-operator-scripts\") pod \"cinder-db-create-9zxn4\" (UID: \"6d9c1c1c-859d-4b2a-9e6b-1a3d863df282\") " pod="openstack/cinder-db-create-9zxn4" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.169733 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj66w\" (UniqueName: \"kubernetes.io/projected/6d9c1c1c-859d-4b2a-9e6b-1a3d863df282-kube-api-access-hj66w\") pod \"cinder-db-create-9zxn4\" (UID: \"6d9c1c1c-859d-4b2a-9e6b-1a3d863df282\") " pod="openstack/cinder-db-create-9zxn4" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.169834 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efe2d244-cf00-49dc-9423-fe757ac195d3-operator-scripts\") pod \"barbican-6b6d-account-create-update-jnz42\" (UID: \"efe2d244-cf00-49dc-9423-fe757ac195d3\") " pod="openstack/barbican-6b6d-account-create-update-jnz42" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.169922 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2xhd\" (UniqueName: \"kubernetes.io/projected/efe2d244-cf00-49dc-9423-fe757ac195d3-kube-api-access-c2xhd\") pod \"barbican-6b6d-account-create-update-jnz42\" (UID: \"efe2d244-cf00-49dc-9423-fe757ac195d3\") " pod="openstack/barbican-6b6d-account-create-update-jnz42" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.213370 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-af0b-account-create-update-wzxbn"] Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.214704 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-af0b-account-create-update-wzxbn" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.217055 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.239107 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-af0b-account-create-update-wzxbn"] Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.271684 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efe2d244-cf00-49dc-9423-fe757ac195d3-operator-scripts\") pod \"barbican-6b6d-account-create-update-jnz42\" (UID: \"efe2d244-cf00-49dc-9423-fe757ac195d3\") " pod="openstack/barbican-6b6d-account-create-update-jnz42" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.271744 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2xhd\" (UniqueName: \"kubernetes.io/projected/efe2d244-cf00-49dc-9423-fe757ac195d3-kube-api-access-c2xhd\") pod \"barbican-6b6d-account-create-update-jnz42\" (UID: \"efe2d244-cf00-49dc-9423-fe757ac195d3\") " pod="openstack/barbican-6b6d-account-create-update-jnz42" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.271843 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7744r\" (UniqueName: \"kubernetes.io/projected/b4a29c27-3526-46bf-a3d2-3aacde054e78-kube-api-access-7744r\") pod \"barbican-db-create-hmdjz\" (UID: \"b4a29c27-3526-46bf-a3d2-3aacde054e78\") " pod="openstack/barbican-db-create-hmdjz" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.271870 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d9c1c1c-859d-4b2a-9e6b-1a3d863df282-operator-scripts\") pod \"cinder-db-create-9zxn4\" (UID: \"6d9c1c1c-859d-4b2a-9e6b-1a3d863df282\") " pod="openstack/cinder-db-create-9zxn4" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.271896 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4a29c27-3526-46bf-a3d2-3aacde054e78-operator-scripts\") pod \"barbican-db-create-hmdjz\" (UID: \"b4a29c27-3526-46bf-a3d2-3aacde054e78\") " pod="openstack/barbican-db-create-hmdjz" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.271961 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj66w\" (UniqueName: \"kubernetes.io/projected/6d9c1c1c-859d-4b2a-9e6b-1a3d863df282-kube-api-access-hj66w\") pod \"cinder-db-create-9zxn4\" (UID: \"6d9c1c1c-859d-4b2a-9e6b-1a3d863df282\") " pod="openstack/cinder-db-create-9zxn4" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.272843 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efe2d244-cf00-49dc-9423-fe757ac195d3-operator-scripts\") pod \"barbican-6b6d-account-create-update-jnz42\" (UID: \"efe2d244-cf00-49dc-9423-fe757ac195d3\") " pod="openstack/barbican-6b6d-account-create-update-jnz42" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.273532 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d9c1c1c-859d-4b2a-9e6b-1a3d863df282-operator-scripts\") pod \"cinder-db-create-9zxn4\" (UID: \"6d9c1c1c-859d-4b2a-9e6b-1a3d863df282\") " pod="openstack/cinder-db-create-9zxn4" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.296874 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2xhd\" (UniqueName: \"kubernetes.io/projected/efe2d244-cf00-49dc-9423-fe757ac195d3-kube-api-access-c2xhd\") pod \"barbican-6b6d-account-create-update-jnz42\" (UID: \"efe2d244-cf00-49dc-9423-fe757ac195d3\") " pod="openstack/barbican-6b6d-account-create-update-jnz42" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.314181 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj66w\" (UniqueName: \"kubernetes.io/projected/6d9c1c1c-859d-4b2a-9e6b-1a3d863df282-kube-api-access-hj66w\") pod \"cinder-db-create-9zxn4\" (UID: \"6d9c1c1c-859d-4b2a-9e6b-1a3d863df282\") " pod="openstack/cinder-db-create-9zxn4" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.374038 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4a29c27-3526-46bf-a3d2-3aacde054e78-operator-scripts\") pod \"barbican-db-create-hmdjz\" (UID: \"b4a29c27-3526-46bf-a3d2-3aacde054e78\") " pod="openstack/barbican-db-create-hmdjz" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.375866 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4a29c27-3526-46bf-a3d2-3aacde054e78-operator-scripts\") pod \"barbican-db-create-hmdjz\" (UID: \"b4a29c27-3526-46bf-a3d2-3aacde054e78\") " pod="openstack/barbican-db-create-hmdjz" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.375927 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/641f63ea-e837-4644-99af-c83045d72395-operator-scripts\") pod \"cinder-af0b-account-create-update-wzxbn\" (UID: \"641f63ea-e837-4644-99af-c83045d72395\") " pod="openstack/cinder-af0b-account-create-update-wzxbn" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.376207 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djm4r\" (UniqueName: \"kubernetes.io/projected/641f63ea-e837-4644-99af-c83045d72395-kube-api-access-djm4r\") pod \"cinder-af0b-account-create-update-wzxbn\" (UID: \"641f63ea-e837-4644-99af-c83045d72395\") " pod="openstack/cinder-af0b-account-create-update-wzxbn" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.376265 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7744r\" (UniqueName: \"kubernetes.io/projected/b4a29c27-3526-46bf-a3d2-3aacde054e78-kube-api-access-7744r\") pod \"barbican-db-create-hmdjz\" (UID: \"b4a29c27-3526-46bf-a3d2-3aacde054e78\") " pod="openstack/barbican-db-create-hmdjz" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.393772 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9zxn4" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.398029 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-fwvm2"] Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.398189 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6b6d-account-create-update-jnz42" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.399309 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7744r\" (UniqueName: \"kubernetes.io/projected/b4a29c27-3526-46bf-a3d2-3aacde054e78-kube-api-access-7744r\") pod \"barbican-db-create-hmdjz\" (UID: \"b4a29c27-3526-46bf-a3d2-3aacde054e78\") " pod="openstack/barbican-db-create-hmdjz" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.399507 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fwvm2" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.410811 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-dnn68"] Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.412787 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dnn68" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.414875 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.415252 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.415477 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.416918 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dxcx7" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.430390 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-hmdjz" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.434115 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-dnn68"] Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.478837 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-fwvm2"] Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.485480 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djm4r\" (UniqueName: \"kubernetes.io/projected/641f63ea-e837-4644-99af-c83045d72395-kube-api-access-djm4r\") pod \"cinder-af0b-account-create-update-wzxbn\" (UID: \"641f63ea-e837-4644-99af-c83045d72395\") " pod="openstack/cinder-af0b-account-create-update-wzxbn" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.485636 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/641f63ea-e837-4644-99af-c83045d72395-operator-scripts\") pod \"cinder-af0b-account-create-update-wzxbn\" (UID: \"641f63ea-e837-4644-99af-c83045d72395\") " pod="openstack/cinder-af0b-account-create-update-wzxbn" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.486474 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/641f63ea-e837-4644-99af-c83045d72395-operator-scripts\") pod \"cinder-af0b-account-create-update-wzxbn\" (UID: \"641f63ea-e837-4644-99af-c83045d72395\") " pod="openstack/cinder-af0b-account-create-update-wzxbn" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.522400 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djm4r\" (UniqueName: \"kubernetes.io/projected/641f63ea-e837-4644-99af-c83045d72395-kube-api-access-djm4r\") pod \"cinder-af0b-account-create-update-wzxbn\" (UID: \"641f63ea-e837-4644-99af-c83045d72395\") " pod="openstack/cinder-af0b-account-create-update-wzxbn" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.540330 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-af0b-account-create-update-wzxbn" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.554613 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d9a-account-create-update-2mbn6"] Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.557827 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9a-account-create-update-2mbn6" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.560439 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.569423 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d9a-account-create-update-2mbn6"] Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.587996 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b30342a-d0e7-4507-8968-05904e98ed9a-config-data\") pod \"keystone-db-sync-dnn68\" (UID: \"7b30342a-d0e7-4507-8968-05904e98ed9a\") " pod="openstack/keystone-db-sync-dnn68" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.588256 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b30342a-d0e7-4507-8968-05904e98ed9a-combined-ca-bundle\") pod \"keystone-db-sync-dnn68\" (UID: \"7b30342a-d0e7-4507-8968-05904e98ed9a\") " pod="openstack/keystone-db-sync-dnn68" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.588406 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2ghk\" (UniqueName: \"kubernetes.io/projected/7b30342a-d0e7-4507-8968-05904e98ed9a-kube-api-access-j2ghk\") pod \"keystone-db-sync-dnn68\" (UID: \"7b30342a-d0e7-4507-8968-05904e98ed9a\") " pod="openstack/keystone-db-sync-dnn68" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.588600 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db6357f7-cfa8-4f21-8b9f-223f828420c3-operator-scripts\") pod \"neutron-db-create-fwvm2\" (UID: \"db6357f7-cfa8-4f21-8b9f-223f828420c3\") " pod="openstack/neutron-db-create-fwvm2" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.588842 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntd9s\" (UniqueName: \"kubernetes.io/projected/db6357f7-cfa8-4f21-8b9f-223f828420c3-kube-api-access-ntd9s\") pod \"neutron-db-create-fwvm2\" (UID: \"db6357f7-cfa8-4f21-8b9f-223f828420c3\") " pod="openstack/neutron-db-create-fwvm2" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.690800 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntd9s\" (UniqueName: \"kubernetes.io/projected/db6357f7-cfa8-4f21-8b9f-223f828420c3-kube-api-access-ntd9s\") pod \"neutron-db-create-fwvm2\" (UID: \"db6357f7-cfa8-4f21-8b9f-223f828420c3\") " pod="openstack/neutron-db-create-fwvm2" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.697450 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b30342a-d0e7-4507-8968-05904e98ed9a-config-data\") pod \"keystone-db-sync-dnn68\" (UID: \"7b30342a-d0e7-4507-8968-05904e98ed9a\") " pod="openstack/keystone-db-sync-dnn68" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.697523 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b30342a-d0e7-4507-8968-05904e98ed9a-combined-ca-bundle\") pod \"keystone-db-sync-dnn68\" (UID: \"7b30342a-d0e7-4507-8968-05904e98ed9a\") " pod="openstack/keystone-db-sync-dnn68" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.697667 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2ghk\" (UniqueName: \"kubernetes.io/projected/7b30342a-d0e7-4507-8968-05904e98ed9a-kube-api-access-j2ghk\") pod \"keystone-db-sync-dnn68\" (UID: \"7b30342a-d0e7-4507-8968-05904e98ed9a\") " pod="openstack/keystone-db-sync-dnn68" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.697738 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db6357f7-cfa8-4f21-8b9f-223f828420c3-operator-scripts\") pod \"neutron-db-create-fwvm2\" (UID: \"db6357f7-cfa8-4f21-8b9f-223f828420c3\") " pod="openstack/neutron-db-create-fwvm2" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.697785 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg2tr\" (UniqueName: \"kubernetes.io/projected/8e6e56df-4807-42c7-a850-4369abb6e61e-kube-api-access-vg2tr\") pod \"neutron-5d9a-account-create-update-2mbn6\" (UID: \"8e6e56df-4807-42c7-a850-4369abb6e61e\") " pod="openstack/neutron-5d9a-account-create-update-2mbn6" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.697861 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e6e56df-4807-42c7-a850-4369abb6e61e-operator-scripts\") pod \"neutron-5d9a-account-create-update-2mbn6\" (UID: \"8e6e56df-4807-42c7-a850-4369abb6e61e\") " pod="openstack/neutron-5d9a-account-create-update-2mbn6" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.699653 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db6357f7-cfa8-4f21-8b9f-223f828420c3-operator-scripts\") pod \"neutron-db-create-fwvm2\" (UID: \"db6357f7-cfa8-4f21-8b9f-223f828420c3\") " pod="openstack/neutron-db-create-fwvm2" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.725374 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b30342a-d0e7-4507-8968-05904e98ed9a-combined-ca-bundle\") pod \"keystone-db-sync-dnn68\" (UID: \"7b30342a-d0e7-4507-8968-05904e98ed9a\") " pod="openstack/keystone-db-sync-dnn68" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.725374 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntd9s\" (UniqueName: \"kubernetes.io/projected/db6357f7-cfa8-4f21-8b9f-223f828420c3-kube-api-access-ntd9s\") pod \"neutron-db-create-fwvm2\" (UID: \"db6357f7-cfa8-4f21-8b9f-223f828420c3\") " pod="openstack/neutron-db-create-fwvm2" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.726031 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2ghk\" (UniqueName: \"kubernetes.io/projected/7b30342a-d0e7-4507-8968-05904e98ed9a-kube-api-access-j2ghk\") pod \"keystone-db-sync-dnn68\" (UID: \"7b30342a-d0e7-4507-8968-05904e98ed9a\") " pod="openstack/keystone-db-sync-dnn68" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.727563 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b30342a-d0e7-4507-8968-05904e98ed9a-config-data\") pod \"keystone-db-sync-dnn68\" (UID: \"7b30342a-d0e7-4507-8968-05904e98ed9a\") " pod="openstack/keystone-db-sync-dnn68" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.798677 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg2tr\" (UniqueName: \"kubernetes.io/projected/8e6e56df-4807-42c7-a850-4369abb6e61e-kube-api-access-vg2tr\") pod \"neutron-5d9a-account-create-update-2mbn6\" (UID: \"8e6e56df-4807-42c7-a850-4369abb6e61e\") " pod="openstack/neutron-5d9a-account-create-update-2mbn6" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.798762 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e6e56df-4807-42c7-a850-4369abb6e61e-operator-scripts\") pod \"neutron-5d9a-account-create-update-2mbn6\" (UID: \"8e6e56df-4807-42c7-a850-4369abb6e61e\") " pod="openstack/neutron-5d9a-account-create-update-2mbn6" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.800060 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e6e56df-4807-42c7-a850-4369abb6e61e-operator-scripts\") pod \"neutron-5d9a-account-create-update-2mbn6\" (UID: \"8e6e56df-4807-42c7-a850-4369abb6e61e\") " pod="openstack/neutron-5d9a-account-create-update-2mbn6" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.816305 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg2tr\" (UniqueName: \"kubernetes.io/projected/8e6e56df-4807-42c7-a850-4369abb6e61e-kube-api-access-vg2tr\") pod \"neutron-5d9a-account-create-update-2mbn6\" (UID: \"8e6e56df-4807-42c7-a850-4369abb6e61e\") " pod="openstack/neutron-5d9a-account-create-update-2mbn6" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.850320 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fwvm2" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.885843 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dnn68" Dec 01 17:42:51 crc kubenswrapper[4868]: I1201 17:42:51.894398 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9a-account-create-update-2mbn6" Dec 01 17:42:52 crc kubenswrapper[4868]: I1201 17:42:52.099140 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6b6d-account-create-update-jnz42"] Dec 01 17:42:52 crc kubenswrapper[4868]: I1201 17:42:52.113810 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-hmdjz"] Dec 01 17:42:52 crc kubenswrapper[4868]: I1201 17:42:52.148175 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-hmdjz" event={"ID":"b4a29c27-3526-46bf-a3d2-3aacde054e78","Type":"ContainerStarted","Data":"d69563754c43787e6020c023fe7f3dfe9e46a3a4288fadb33de43848aec20db5"} Dec 01 17:42:52 crc kubenswrapper[4868]: I1201 17:42:52.149081 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6b6d-account-create-update-jnz42" event={"ID":"efe2d244-cf00-49dc-9423-fe757ac195d3","Type":"ContainerStarted","Data":"3afe445c7b2d7f932415a2cbfd5f1adcc1f6c253e9054ed873f33e4821264c5d"} Dec 01 17:42:52 crc kubenswrapper[4868]: I1201 17:42:52.203192 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-fwvm2"] Dec 01 17:42:52 crc kubenswrapper[4868]: I1201 17:42:52.243925 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-af0b-account-create-update-wzxbn"] Dec 01 17:42:52 crc kubenswrapper[4868]: I1201 17:42:52.259324 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-9zxn4"] Dec 01 17:42:52 crc kubenswrapper[4868]: W1201 17:42:52.270187 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d9c1c1c_859d_4b2a_9e6b_1a3d863df282.slice/crio-ab6add83c572e320b5e4e6274cc712a1140038f30fb4b1b1dcf35d8965c80baa WatchSource:0}: Error finding container ab6add83c572e320b5e4e6274cc712a1140038f30fb4b1b1dcf35d8965c80baa: Status 404 returned error can't find the container with id ab6add83c572e320b5e4e6274cc712a1140038f30fb4b1b1dcf35d8965c80baa Dec 01 17:42:52 crc kubenswrapper[4868]: I1201 17:42:52.507795 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-dnn68"] Dec 01 17:42:52 crc kubenswrapper[4868]: I1201 17:42:52.517263 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d9a-account-create-update-2mbn6"] Dec 01 17:42:52 crc kubenswrapper[4868]: W1201 17:42:52.521429 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b30342a_d0e7_4507_8968_05904e98ed9a.slice/crio-fe043f36d859f5b7130b2973c803b672defb7de5ba0097c33b73092d93e9d173 WatchSource:0}: Error finding container fe043f36d859f5b7130b2973c803b672defb7de5ba0097c33b73092d93e9d173: Status 404 returned error can't find the container with id fe043f36d859f5b7130b2973c803b672defb7de5ba0097c33b73092d93e9d173 Dec 01 17:42:52 crc kubenswrapper[4868]: W1201 17:42:52.542922 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e6e56df_4807_42c7_a850_4369abb6e61e.slice/crio-08b6cd6c6cadfc3a601b2fe21115f45717ab73be64aa6c1d32b8b4fb4ea9c51c WatchSource:0}: Error finding container 08b6cd6c6cadfc3a601b2fe21115f45717ab73be64aa6c1d32b8b4fb4ea9c51c: Status 404 returned error can't find the container with id 08b6cd6c6cadfc3a601b2fe21115f45717ab73be64aa6c1d32b8b4fb4ea9c51c Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.159417 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9a-account-create-update-2mbn6" event={"ID":"8e6e56df-4807-42c7-a850-4369abb6e61e","Type":"ContainerStarted","Data":"a3eddc1407dda99e2d67d97f23e6008b027a9ac998e0680735a4a33f4d37628c"} Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.159767 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9a-account-create-update-2mbn6" event={"ID":"8e6e56df-4807-42c7-a850-4369abb6e61e","Type":"ContainerStarted","Data":"08b6cd6c6cadfc3a601b2fe21115f45717ab73be64aa6c1d32b8b4fb4ea9c51c"} Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.160918 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-af0b-account-create-update-wzxbn" event={"ID":"641f63ea-e837-4644-99af-c83045d72395","Type":"ContainerStarted","Data":"ff5c1ea23acc346cf2d00fa44c161aa13ee4092a071208bfce781277cc369d92"} Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.161001 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-af0b-account-create-update-wzxbn" event={"ID":"641f63ea-e837-4644-99af-c83045d72395","Type":"ContainerStarted","Data":"0b29408c76d4851c533ea51c6ba4e0b6b4099ce0f08b87763a06902b9b47cd2c"} Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.162282 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dnn68" event={"ID":"7b30342a-d0e7-4507-8968-05904e98ed9a","Type":"ContainerStarted","Data":"fe043f36d859f5b7130b2973c803b672defb7de5ba0097c33b73092d93e9d173"} Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.164575 4868 generic.go:334] "Generic (PLEG): container finished" podID="b4a29c27-3526-46bf-a3d2-3aacde054e78" containerID="c70a918b6b5630a265c0fdfb8dbc21f7c478a51f339e0f48254f56d912e7a052" exitCode=0 Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.164651 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-hmdjz" event={"ID":"b4a29c27-3526-46bf-a3d2-3aacde054e78","Type":"ContainerDied","Data":"c70a918b6b5630a265c0fdfb8dbc21f7c478a51f339e0f48254f56d912e7a052"} Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.166202 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9zxn4" event={"ID":"6d9c1c1c-859d-4b2a-9e6b-1a3d863df282","Type":"ContainerStarted","Data":"cd7e61c92876982a7e924490b9dd7414a5c021c25911071f3ac28d7684e0740e"} Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.166230 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9zxn4" event={"ID":"6d9c1c1c-859d-4b2a-9e6b-1a3d863df282","Type":"ContainerStarted","Data":"ab6add83c572e320b5e4e6274cc712a1140038f30fb4b1b1dcf35d8965c80baa"} Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.168572 4868 generic.go:334] "Generic (PLEG): container finished" podID="efe2d244-cf00-49dc-9423-fe757ac195d3" containerID="91d07288ef7ac072439586ff0d732c58771e82e82f3785f5f153b621bfff759b" exitCode=0 Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.168637 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6b6d-account-create-update-jnz42" event={"ID":"efe2d244-cf00-49dc-9423-fe757ac195d3","Type":"ContainerDied","Data":"91d07288ef7ac072439586ff0d732c58771e82e82f3785f5f153b621bfff759b"} Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.170612 4868 generic.go:334] "Generic (PLEG): container finished" podID="db6357f7-cfa8-4f21-8b9f-223f828420c3" containerID="bb683e475676090d0da843b669bb2d444dd1c46715f86f31e45d7d5d91c5e3ba" exitCode=0 Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.170631 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fwvm2" event={"ID":"db6357f7-cfa8-4f21-8b9f-223f828420c3","Type":"ContainerDied","Data":"bb683e475676090d0da843b669bb2d444dd1c46715f86f31e45d7d5d91c5e3ba"} Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.170657 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fwvm2" event={"ID":"db6357f7-cfa8-4f21-8b9f-223f828420c3","Type":"ContainerStarted","Data":"e13ebe6a34a55cdcccbc1484d704a1aea8665a620f8c4ccc49153ceea2c79ebe"} Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.189017 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d9a-account-create-update-2mbn6" podStartSLOduration=2.188998093 podStartE2EDuration="2.188998093s" podCreationTimestamp="2025-12-01 17:42:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:42:53.185283618 +0000 UTC m=+1045.556394029" watchObservedRunningTime="2025-12-01 17:42:53.188998093 +0000 UTC m=+1045.560108504" Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.268855 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-af0b-account-create-update-wzxbn" podStartSLOduration=2.268833553 podStartE2EDuration="2.268833553s" podCreationTimestamp="2025-12-01 17:42:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:42:53.266326009 +0000 UTC m=+1045.637436420" watchObservedRunningTime="2025-12-01 17:42:53.268833553 +0000 UTC m=+1045.639943964" Dec 01 17:42:53 crc kubenswrapper[4868]: I1201 17:42:53.293307 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-9zxn4" podStartSLOduration=3.293282385 podStartE2EDuration="3.293282385s" podCreationTimestamp="2025-12-01 17:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:42:53.287825146 +0000 UTC m=+1045.658935557" watchObservedRunningTime="2025-12-01 17:42:53.293282385 +0000 UTC m=+1045.664392796" Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.204647 4868 generic.go:334] "Generic (PLEG): container finished" podID="641f63ea-e837-4644-99af-c83045d72395" containerID="ff5c1ea23acc346cf2d00fa44c161aa13ee4092a071208bfce781277cc369d92" exitCode=0 Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.204877 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-af0b-account-create-update-wzxbn" event={"ID":"641f63ea-e837-4644-99af-c83045d72395","Type":"ContainerDied","Data":"ff5c1ea23acc346cf2d00fa44c161aa13ee4092a071208bfce781277cc369d92"} Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.210011 4868 generic.go:334] "Generic (PLEG): container finished" podID="6d9c1c1c-859d-4b2a-9e6b-1a3d863df282" containerID="cd7e61c92876982a7e924490b9dd7414a5c021c25911071f3ac28d7684e0740e" exitCode=0 Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.210075 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9zxn4" event={"ID":"6d9c1c1c-859d-4b2a-9e6b-1a3d863df282","Type":"ContainerDied","Data":"cd7e61c92876982a7e924490b9dd7414a5c021c25911071f3ac28d7684e0740e"} Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.212430 4868 generic.go:334] "Generic (PLEG): container finished" podID="8e6e56df-4807-42c7-a850-4369abb6e61e" containerID="a3eddc1407dda99e2d67d97f23e6008b027a9ac998e0680735a4a33f4d37628c" exitCode=0 Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.212509 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9a-account-create-update-2mbn6" event={"ID":"8e6e56df-4807-42c7-a850-4369abb6e61e","Type":"ContainerDied","Data":"a3eddc1407dda99e2d67d97f23e6008b027a9ac998e0680735a4a33f4d37628c"} Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.708585 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fwvm2" Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.724834 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-hmdjz" Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.730370 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6b6d-account-create-update-jnz42" Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.855153 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntd9s\" (UniqueName: \"kubernetes.io/projected/db6357f7-cfa8-4f21-8b9f-223f828420c3-kube-api-access-ntd9s\") pod \"db6357f7-cfa8-4f21-8b9f-223f828420c3\" (UID: \"db6357f7-cfa8-4f21-8b9f-223f828420c3\") " Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.855672 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efe2d244-cf00-49dc-9423-fe757ac195d3-operator-scripts\") pod \"efe2d244-cf00-49dc-9423-fe757ac195d3\" (UID: \"efe2d244-cf00-49dc-9423-fe757ac195d3\") " Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.855754 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db6357f7-cfa8-4f21-8b9f-223f828420c3-operator-scripts\") pod \"db6357f7-cfa8-4f21-8b9f-223f828420c3\" (UID: \"db6357f7-cfa8-4f21-8b9f-223f828420c3\") " Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.855777 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7744r\" (UniqueName: \"kubernetes.io/projected/b4a29c27-3526-46bf-a3d2-3aacde054e78-kube-api-access-7744r\") pod \"b4a29c27-3526-46bf-a3d2-3aacde054e78\" (UID: \"b4a29c27-3526-46bf-a3d2-3aacde054e78\") " Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.855806 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2xhd\" (UniqueName: \"kubernetes.io/projected/efe2d244-cf00-49dc-9423-fe757ac195d3-kube-api-access-c2xhd\") pod \"efe2d244-cf00-49dc-9423-fe757ac195d3\" (UID: \"efe2d244-cf00-49dc-9423-fe757ac195d3\") " Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.855919 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4a29c27-3526-46bf-a3d2-3aacde054e78-operator-scripts\") pod \"b4a29c27-3526-46bf-a3d2-3aacde054e78\" (UID: \"b4a29c27-3526-46bf-a3d2-3aacde054e78\") " Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.856368 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db6357f7-cfa8-4f21-8b9f-223f828420c3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "db6357f7-cfa8-4f21-8b9f-223f828420c3" (UID: "db6357f7-cfa8-4f21-8b9f-223f828420c3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.856474 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efe2d244-cf00-49dc-9423-fe757ac195d3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "efe2d244-cf00-49dc-9423-fe757ac195d3" (UID: "efe2d244-cf00-49dc-9423-fe757ac195d3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.856739 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4a29c27-3526-46bf-a3d2-3aacde054e78-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b4a29c27-3526-46bf-a3d2-3aacde054e78" (UID: "b4a29c27-3526-46bf-a3d2-3aacde054e78"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.863175 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efe2d244-cf00-49dc-9423-fe757ac195d3-kube-api-access-c2xhd" (OuterVolumeSpecName: "kube-api-access-c2xhd") pod "efe2d244-cf00-49dc-9423-fe757ac195d3" (UID: "efe2d244-cf00-49dc-9423-fe757ac195d3"). InnerVolumeSpecName "kube-api-access-c2xhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.864607 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4a29c27-3526-46bf-a3d2-3aacde054e78-kube-api-access-7744r" (OuterVolumeSpecName: "kube-api-access-7744r") pod "b4a29c27-3526-46bf-a3d2-3aacde054e78" (UID: "b4a29c27-3526-46bf-a3d2-3aacde054e78"). InnerVolumeSpecName "kube-api-access-7744r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.864726 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db6357f7-cfa8-4f21-8b9f-223f828420c3-kube-api-access-ntd9s" (OuterVolumeSpecName: "kube-api-access-ntd9s") pod "db6357f7-cfa8-4f21-8b9f-223f828420c3" (UID: "db6357f7-cfa8-4f21-8b9f-223f828420c3"). InnerVolumeSpecName "kube-api-access-ntd9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.962564 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efe2d244-cf00-49dc-9423-fe757ac195d3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.962601 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db6357f7-cfa8-4f21-8b9f-223f828420c3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.962611 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7744r\" (UniqueName: \"kubernetes.io/projected/b4a29c27-3526-46bf-a3d2-3aacde054e78-kube-api-access-7744r\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.962621 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2xhd\" (UniqueName: \"kubernetes.io/projected/efe2d244-cf00-49dc-9423-fe757ac195d3-kube-api-access-c2xhd\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.962631 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4a29c27-3526-46bf-a3d2-3aacde054e78-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:54 crc kubenswrapper[4868]: I1201 17:42:54.962639 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntd9s\" (UniqueName: \"kubernetes.io/projected/db6357f7-cfa8-4f21-8b9f-223f828420c3-kube-api-access-ntd9s\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:55 crc kubenswrapper[4868]: I1201 17:42:55.224644 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-hmdjz" event={"ID":"b4a29c27-3526-46bf-a3d2-3aacde054e78","Type":"ContainerDied","Data":"d69563754c43787e6020c023fe7f3dfe9e46a3a4288fadb33de43848aec20db5"} Dec 01 17:42:55 crc kubenswrapper[4868]: I1201 17:42:55.224709 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d69563754c43787e6020c023fe7f3dfe9e46a3a4288fadb33de43848aec20db5" Dec 01 17:42:55 crc kubenswrapper[4868]: I1201 17:42:55.224674 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-hmdjz" Dec 01 17:42:55 crc kubenswrapper[4868]: I1201 17:42:55.226524 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6b6d-account-create-update-jnz42" Dec 01 17:42:55 crc kubenswrapper[4868]: I1201 17:42:55.226700 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6b6d-account-create-update-jnz42" event={"ID":"efe2d244-cf00-49dc-9423-fe757ac195d3","Type":"ContainerDied","Data":"3afe445c7b2d7f932415a2cbfd5f1adcc1f6c253e9054ed873f33e4821264c5d"} Dec 01 17:42:55 crc kubenswrapper[4868]: I1201 17:42:55.226738 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3afe445c7b2d7f932415a2cbfd5f1adcc1f6c253e9054ed873f33e4821264c5d" Dec 01 17:42:55 crc kubenswrapper[4868]: I1201 17:42:55.228539 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fwvm2" event={"ID":"db6357f7-cfa8-4f21-8b9f-223f828420c3","Type":"ContainerDied","Data":"e13ebe6a34a55cdcccbc1484d704a1aea8665a620f8c4ccc49153ceea2c79ebe"} Dec 01 17:42:55 crc kubenswrapper[4868]: I1201 17:42:55.228589 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e13ebe6a34a55cdcccbc1484d704a1aea8665a620f8c4ccc49153ceea2c79ebe" Dec 01 17:42:55 crc kubenswrapper[4868]: I1201 17:42:55.228659 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fwvm2" Dec 01 17:42:55 crc kubenswrapper[4868]: I1201 17:42:55.905303 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:42:55 crc kubenswrapper[4868]: I1201 17:42:55.905680 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:42:55 crc kubenswrapper[4868]: I1201 17:42:55.905764 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:42:55 crc kubenswrapper[4868]: I1201 17:42:55.907031 4868 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2d56a7ef7fc39bd11e02dfcafda9203c34f3593dd8178f744951427000e66a77"} pod="openshift-machine-config-operator/machine-config-daemon-njgpn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 17:42:55 crc kubenswrapper[4868]: I1201 17:42:55.907102 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" containerID="cri-o://2d56a7ef7fc39bd11e02dfcafda9203c34f3593dd8178f744951427000e66a77" gracePeriod=600 Dec 01 17:42:56 crc kubenswrapper[4868]: I1201 17:42:56.239982 4868 generic.go:334] "Generic (PLEG): container finished" podID="ec4e08bf-df56-4490-8e79-a4175c727405" containerID="2d56a7ef7fc39bd11e02dfcafda9203c34f3593dd8178f744951427000e66a77" exitCode=0 Dec 01 17:42:56 crc kubenswrapper[4868]: I1201 17:42:56.240027 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerDied","Data":"2d56a7ef7fc39bd11e02dfcafda9203c34f3593dd8178f744951427000e66a77"} Dec 01 17:42:56 crc kubenswrapper[4868]: I1201 17:42:56.240063 4868 scope.go:117] "RemoveContainer" containerID="5bcb6e04848aace81d63d47c22db36f1a32f6313299199f19533a292326b33e3" Dec 01 17:42:56 crc kubenswrapper[4868]: I1201 17:42:56.890984 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:42:56 crc kubenswrapper[4868]: I1201 17:42:56.960632 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-tcxlc"] Dec 01 17:42:56 crc kubenswrapper[4868]: I1201 17:42:56.960965 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" podUID="27bdb46a-6c28-4f18-82e7-50d25471969c" containerName="dnsmasq-dns" containerID="cri-o://d68162bfc8404f519b834a7cb4da209f50f4de3ecef416f60b6be3e708c34a31" gracePeriod=10 Dec 01 17:42:57 crc kubenswrapper[4868]: I1201 17:42:57.254078 4868 generic.go:334] "Generic (PLEG): container finished" podID="27bdb46a-6c28-4f18-82e7-50d25471969c" containerID="d68162bfc8404f519b834a7cb4da209f50f4de3ecef416f60b6be3e708c34a31" exitCode=0 Dec 01 17:42:57 crc kubenswrapper[4868]: I1201 17:42:57.254417 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" event={"ID":"27bdb46a-6c28-4f18-82e7-50d25471969c","Type":"ContainerDied","Data":"d68162bfc8404f519b834a7cb4da209f50f4de3ecef416f60b6be3e708c34a31"} Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.245580 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-af0b-account-create-update-wzxbn" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.269163 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9zxn4" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.271668 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9a-account-create-update-2mbn6" event={"ID":"8e6e56df-4807-42c7-a850-4369abb6e61e","Type":"ContainerDied","Data":"08b6cd6c6cadfc3a601b2fe21115f45717ab73be64aa6c1d32b8b4fb4ea9c51c"} Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.271720 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08b6cd6c6cadfc3a601b2fe21115f45717ab73be64aa6c1d32b8b4fb4ea9c51c" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.283623 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-af0b-account-create-update-wzxbn" event={"ID":"641f63ea-e837-4644-99af-c83045d72395","Type":"ContainerDied","Data":"0b29408c76d4851c533ea51c6ba4e0b6b4099ce0f08b87763a06902b9b47cd2c"} Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.283705 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b29408c76d4851c533ea51c6ba4e0b6b4099ce0f08b87763a06902b9b47cd2c" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.283808 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-af0b-account-create-update-wzxbn" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.309383 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9zxn4" event={"ID":"6d9c1c1c-859d-4b2a-9e6b-1a3d863df282","Type":"ContainerDied","Data":"ab6add83c572e320b5e4e6274cc712a1140038f30fb4b1b1dcf35d8965c80baa"} Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.309440 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab6add83c572e320b5e4e6274cc712a1140038f30fb4b1b1dcf35d8965c80baa" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.309516 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9zxn4" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.360740 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/641f63ea-e837-4644-99af-c83045d72395-operator-scripts\") pod \"641f63ea-e837-4644-99af-c83045d72395\" (UID: \"641f63ea-e837-4644-99af-c83045d72395\") " Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.360833 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d9c1c1c-859d-4b2a-9e6b-1a3d863df282-operator-scripts\") pod \"6d9c1c1c-859d-4b2a-9e6b-1a3d863df282\" (UID: \"6d9c1c1c-859d-4b2a-9e6b-1a3d863df282\") " Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.360890 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djm4r\" (UniqueName: \"kubernetes.io/projected/641f63ea-e837-4644-99af-c83045d72395-kube-api-access-djm4r\") pod \"641f63ea-e837-4644-99af-c83045d72395\" (UID: \"641f63ea-e837-4644-99af-c83045d72395\") " Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.360966 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj66w\" (UniqueName: \"kubernetes.io/projected/6d9c1c1c-859d-4b2a-9e6b-1a3d863df282-kube-api-access-hj66w\") pod \"6d9c1c1c-859d-4b2a-9e6b-1a3d863df282\" (UID: \"6d9c1c1c-859d-4b2a-9e6b-1a3d863df282\") " Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.363320 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d9c1c1c-859d-4b2a-9e6b-1a3d863df282-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6d9c1c1c-859d-4b2a-9e6b-1a3d863df282" (UID: "6d9c1c1c-859d-4b2a-9e6b-1a3d863df282"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.363370 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/641f63ea-e837-4644-99af-c83045d72395-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "641f63ea-e837-4644-99af-c83045d72395" (UID: "641f63ea-e837-4644-99af-c83045d72395"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.364926 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/641f63ea-e837-4644-99af-c83045d72395-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.365080 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d9c1c1c-859d-4b2a-9e6b-1a3d863df282-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.366439 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9a-account-create-update-2mbn6" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.369996 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d9c1c1c-859d-4b2a-9e6b-1a3d863df282-kube-api-access-hj66w" (OuterVolumeSpecName: "kube-api-access-hj66w") pod "6d9c1c1c-859d-4b2a-9e6b-1a3d863df282" (UID: "6d9c1c1c-859d-4b2a-9e6b-1a3d863df282"). InnerVolumeSpecName "kube-api-access-hj66w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.370772 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/641f63ea-e837-4644-99af-c83045d72395-kube-api-access-djm4r" (OuterVolumeSpecName: "kube-api-access-djm4r") pod "641f63ea-e837-4644-99af-c83045d72395" (UID: "641f63ea-e837-4644-99af-c83045d72395"). InnerVolumeSpecName "kube-api-access-djm4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.375657 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.466019 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-ovsdbserver-sb\") pod \"27bdb46a-6c28-4f18-82e7-50d25471969c\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.466371 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vg2tr\" (UniqueName: \"kubernetes.io/projected/8e6e56df-4807-42c7-a850-4369abb6e61e-kube-api-access-vg2tr\") pod \"8e6e56df-4807-42c7-a850-4369abb6e61e\" (UID: \"8e6e56df-4807-42c7-a850-4369abb6e61e\") " Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.466470 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-ovsdbserver-nb\") pod \"27bdb46a-6c28-4f18-82e7-50d25471969c\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.466495 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-config\") pod \"27bdb46a-6c28-4f18-82e7-50d25471969c\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.466619 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e6e56df-4807-42c7-a850-4369abb6e61e-operator-scripts\") pod \"8e6e56df-4807-42c7-a850-4369abb6e61e\" (UID: \"8e6e56df-4807-42c7-a850-4369abb6e61e\") " Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.466638 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdv5k\" (UniqueName: \"kubernetes.io/projected/27bdb46a-6c28-4f18-82e7-50d25471969c-kube-api-access-wdv5k\") pod \"27bdb46a-6c28-4f18-82e7-50d25471969c\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.466673 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-dns-svc\") pod \"27bdb46a-6c28-4f18-82e7-50d25471969c\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.466988 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj66w\" (UniqueName: \"kubernetes.io/projected/6d9c1c1c-859d-4b2a-9e6b-1a3d863df282-kube-api-access-hj66w\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.467003 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djm4r\" (UniqueName: \"kubernetes.io/projected/641f63ea-e837-4644-99af-c83045d72395-kube-api-access-djm4r\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.472273 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e6e56df-4807-42c7-a850-4369abb6e61e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8e6e56df-4807-42c7-a850-4369abb6e61e" (UID: "8e6e56df-4807-42c7-a850-4369abb6e61e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.483966 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27bdb46a-6c28-4f18-82e7-50d25471969c-kube-api-access-wdv5k" (OuterVolumeSpecName: "kube-api-access-wdv5k") pod "27bdb46a-6c28-4f18-82e7-50d25471969c" (UID: "27bdb46a-6c28-4f18-82e7-50d25471969c"). InnerVolumeSpecName "kube-api-access-wdv5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.484112 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e6e56df-4807-42c7-a850-4369abb6e61e-kube-api-access-vg2tr" (OuterVolumeSpecName: "kube-api-access-vg2tr") pod "8e6e56df-4807-42c7-a850-4369abb6e61e" (UID: "8e6e56df-4807-42c7-a850-4369abb6e61e"). InnerVolumeSpecName "kube-api-access-vg2tr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.519087 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "27bdb46a-6c28-4f18-82e7-50d25471969c" (UID: "27bdb46a-6c28-4f18-82e7-50d25471969c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.524747 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "27bdb46a-6c28-4f18-82e7-50d25471969c" (UID: "27bdb46a-6c28-4f18-82e7-50d25471969c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:58 crc kubenswrapper[4868]: E1201 17:42:58.527790 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-config podName:27bdb46a-6c28-4f18-82e7-50d25471969c nodeName:}" failed. No retries permitted until 2025-12-01 17:42:59.027734192 +0000 UTC m=+1051.398844603 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-config") pod "27bdb46a-6c28-4f18-82e7-50d25471969c" (UID: "27bdb46a-6c28-4f18-82e7-50d25471969c") : error deleting /var/lib/kubelet/pods/27bdb46a-6c28-4f18-82e7-50d25471969c/volume-subpaths: remove /var/lib/kubelet/pods/27bdb46a-6c28-4f18-82e7-50d25471969c/volume-subpaths: no such file or directory Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.528388 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "27bdb46a-6c28-4f18-82e7-50d25471969c" (UID: "27bdb46a-6c28-4f18-82e7-50d25471969c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.568883 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e6e56df-4807-42c7-a850-4369abb6e61e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.568925 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdv5k\" (UniqueName: \"kubernetes.io/projected/27bdb46a-6c28-4f18-82e7-50d25471969c-kube-api-access-wdv5k\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.568954 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.568963 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.568972 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vg2tr\" (UniqueName: \"kubernetes.io/projected/8e6e56df-4807-42c7-a850-4369abb6e61e-kube-api-access-vg2tr\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:58 crc kubenswrapper[4868]: I1201 17:42:58.568981 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:59 crc kubenswrapper[4868]: I1201 17:42:59.077040 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-config\") pod \"27bdb46a-6c28-4f18-82e7-50d25471969c\" (UID: \"27bdb46a-6c28-4f18-82e7-50d25471969c\") " Dec 01 17:42:59 crc kubenswrapper[4868]: I1201 17:42:59.077968 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-config" (OuterVolumeSpecName: "config") pod "27bdb46a-6c28-4f18-82e7-50d25471969c" (UID: "27bdb46a-6c28-4f18-82e7-50d25471969c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:42:59 crc kubenswrapper[4868]: I1201 17:42:59.180297 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27bdb46a-6c28-4f18-82e7-50d25471969c-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:42:59 crc kubenswrapper[4868]: I1201 17:42:59.320262 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"35d0f81eb6527dda2c61338bf9c12106c16489b5c045a71ea474df12051c7650"} Dec 01 17:42:59 crc kubenswrapper[4868]: I1201 17:42:59.323167 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dnn68" event={"ID":"7b30342a-d0e7-4507-8968-05904e98ed9a","Type":"ContainerStarted","Data":"45963812b655d654c05a615cbeb3bddfff82bba9799828eea2cbdd413b56f694"} Dec 01 17:42:59 crc kubenswrapper[4868]: I1201 17:42:59.326073 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9a-account-create-update-2mbn6" Dec 01 17:42:59 crc kubenswrapper[4868]: I1201 17:42:59.326058 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" event={"ID":"27bdb46a-6c28-4f18-82e7-50d25471969c","Type":"ContainerDied","Data":"bf7c17c03e968e3a6522cada2e58d9eff3791106b6141f6d1ff4434e1f90f029"} Dec 01 17:42:59 crc kubenswrapper[4868]: I1201 17:42:59.326258 4868 scope.go:117] "RemoveContainer" containerID="d68162bfc8404f519b834a7cb4da209f50f4de3ecef416f60b6be3e708c34a31" Dec 01 17:42:59 crc kubenswrapper[4868]: I1201 17:42:59.326600 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-tcxlc" Dec 01 17:42:59 crc kubenswrapper[4868]: I1201 17:42:59.369243 4868 scope.go:117] "RemoveContainer" containerID="3f079fdf0fdedea58f7a4baae3764b6eae28f2a934a6223b357708c6283f6c73" Dec 01 17:42:59 crc kubenswrapper[4868]: I1201 17:42:59.447164 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-dnn68" podStartSLOduration=2.894370683 podStartE2EDuration="8.447141856s" podCreationTimestamp="2025-12-01 17:42:51 +0000 UTC" firstStartedPulling="2025-12-01 17:42:52.523917618 +0000 UTC m=+1044.895028029" lastFinishedPulling="2025-12-01 17:42:58.076688791 +0000 UTC m=+1050.447799202" observedRunningTime="2025-12-01 17:42:59.408564925 +0000 UTC m=+1051.779675326" watchObservedRunningTime="2025-12-01 17:42:59.447141856 +0000 UTC m=+1051.818252267" Dec 01 17:42:59 crc kubenswrapper[4868]: I1201 17:42:59.482820 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-tcxlc"] Dec 01 17:42:59 crc kubenswrapper[4868]: I1201 17:42:59.488208 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-tcxlc"] Dec 01 17:43:00 crc kubenswrapper[4868]: I1201 17:43:00.183125 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27bdb46a-6c28-4f18-82e7-50d25471969c" path="/var/lib/kubelet/pods/27bdb46a-6c28-4f18-82e7-50d25471969c/volumes" Dec 01 17:43:01 crc kubenswrapper[4868]: I1201 17:43:01.354062 4868 generic.go:334] "Generic (PLEG): container finished" podID="7b30342a-d0e7-4507-8968-05904e98ed9a" containerID="45963812b655d654c05a615cbeb3bddfff82bba9799828eea2cbdd413b56f694" exitCode=0 Dec 01 17:43:01 crc kubenswrapper[4868]: I1201 17:43:01.354145 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dnn68" event={"ID":"7b30342a-d0e7-4507-8968-05904e98ed9a","Type":"ContainerDied","Data":"45963812b655d654c05a615cbeb3bddfff82bba9799828eea2cbdd413b56f694"} Dec 01 17:43:02 crc kubenswrapper[4868]: I1201 17:43:02.703366 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dnn68" Dec 01 17:43:02 crc kubenswrapper[4868]: I1201 17:43:02.750571 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b30342a-d0e7-4507-8968-05904e98ed9a-combined-ca-bundle\") pod \"7b30342a-d0e7-4507-8968-05904e98ed9a\" (UID: \"7b30342a-d0e7-4507-8968-05904e98ed9a\") " Dec 01 17:43:02 crc kubenswrapper[4868]: I1201 17:43:02.750717 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2ghk\" (UniqueName: \"kubernetes.io/projected/7b30342a-d0e7-4507-8968-05904e98ed9a-kube-api-access-j2ghk\") pod \"7b30342a-d0e7-4507-8968-05904e98ed9a\" (UID: \"7b30342a-d0e7-4507-8968-05904e98ed9a\") " Dec 01 17:43:02 crc kubenswrapper[4868]: I1201 17:43:02.750772 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b30342a-d0e7-4507-8968-05904e98ed9a-config-data\") pod \"7b30342a-d0e7-4507-8968-05904e98ed9a\" (UID: \"7b30342a-d0e7-4507-8968-05904e98ed9a\") " Dec 01 17:43:02 crc kubenswrapper[4868]: I1201 17:43:02.760254 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b30342a-d0e7-4507-8968-05904e98ed9a-kube-api-access-j2ghk" (OuterVolumeSpecName: "kube-api-access-j2ghk") pod "7b30342a-d0e7-4507-8968-05904e98ed9a" (UID: "7b30342a-d0e7-4507-8968-05904e98ed9a"). InnerVolumeSpecName "kube-api-access-j2ghk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:43:02 crc kubenswrapper[4868]: I1201 17:43:02.787658 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b30342a-d0e7-4507-8968-05904e98ed9a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b30342a-d0e7-4507-8968-05904e98ed9a" (UID: "7b30342a-d0e7-4507-8968-05904e98ed9a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:02 crc kubenswrapper[4868]: I1201 17:43:02.800070 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b30342a-d0e7-4507-8968-05904e98ed9a-config-data" (OuterVolumeSpecName: "config-data") pod "7b30342a-d0e7-4507-8968-05904e98ed9a" (UID: "7b30342a-d0e7-4507-8968-05904e98ed9a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:02 crc kubenswrapper[4868]: I1201 17:43:02.852624 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b30342a-d0e7-4507-8968-05904e98ed9a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:02 crc kubenswrapper[4868]: I1201 17:43:02.852665 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2ghk\" (UniqueName: \"kubernetes.io/projected/7b30342a-d0e7-4507-8968-05904e98ed9a-kube-api-access-j2ghk\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:02 crc kubenswrapper[4868]: I1201 17:43:02.852699 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b30342a-d0e7-4507-8968-05904e98ed9a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.373575 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dnn68" event={"ID":"7b30342a-d0e7-4507-8968-05904e98ed9a","Type":"ContainerDied","Data":"fe043f36d859f5b7130b2973c803b672defb7de5ba0097c33b73092d93e9d173"} Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.373630 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe043f36d859f5b7130b2973c803b672defb7de5ba0097c33b73092d93e9d173" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.373650 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dnn68" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.989730 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-6lddf"] Dec 01 17:43:03 crc kubenswrapper[4868]: E1201 17:43:03.992603 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db6357f7-cfa8-4f21-8b9f-223f828420c3" containerName="mariadb-database-create" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.992621 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="db6357f7-cfa8-4f21-8b9f-223f828420c3" containerName="mariadb-database-create" Dec 01 17:43:03 crc kubenswrapper[4868]: E1201 17:43:03.992631 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6e56df-4807-42c7-a850-4369abb6e61e" containerName="mariadb-account-create-update" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.992636 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6e56df-4807-42c7-a850-4369abb6e61e" containerName="mariadb-account-create-update" Dec 01 17:43:03 crc kubenswrapper[4868]: E1201 17:43:03.992664 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27bdb46a-6c28-4f18-82e7-50d25471969c" containerName="dnsmasq-dns" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.992672 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="27bdb46a-6c28-4f18-82e7-50d25471969c" containerName="dnsmasq-dns" Dec 01 17:43:03 crc kubenswrapper[4868]: E1201 17:43:03.992686 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efe2d244-cf00-49dc-9423-fe757ac195d3" containerName="mariadb-account-create-update" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.992692 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="efe2d244-cf00-49dc-9423-fe757ac195d3" containerName="mariadb-account-create-update" Dec 01 17:43:03 crc kubenswrapper[4868]: E1201 17:43:03.992703 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4a29c27-3526-46bf-a3d2-3aacde054e78" containerName="mariadb-database-create" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.992709 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4a29c27-3526-46bf-a3d2-3aacde054e78" containerName="mariadb-database-create" Dec 01 17:43:03 crc kubenswrapper[4868]: E1201 17:43:03.992719 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d9c1c1c-859d-4b2a-9e6b-1a3d863df282" containerName="mariadb-database-create" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.992725 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d9c1c1c-859d-4b2a-9e6b-1a3d863df282" containerName="mariadb-database-create" Dec 01 17:43:03 crc kubenswrapper[4868]: E1201 17:43:03.992733 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27bdb46a-6c28-4f18-82e7-50d25471969c" containerName="init" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.992739 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="27bdb46a-6c28-4f18-82e7-50d25471969c" containerName="init" Dec 01 17:43:03 crc kubenswrapper[4868]: E1201 17:43:03.992748 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b30342a-d0e7-4507-8968-05904e98ed9a" containerName="keystone-db-sync" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.992754 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b30342a-d0e7-4507-8968-05904e98ed9a" containerName="keystone-db-sync" Dec 01 17:43:03 crc kubenswrapper[4868]: E1201 17:43:03.992764 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="641f63ea-e837-4644-99af-c83045d72395" containerName="mariadb-account-create-update" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.992771 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="641f63ea-e837-4644-99af-c83045d72395" containerName="mariadb-account-create-update" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.992935 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b30342a-d0e7-4507-8968-05904e98ed9a" containerName="keystone-db-sync" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.992964 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="efe2d244-cf00-49dc-9423-fe757ac195d3" containerName="mariadb-account-create-update" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.992977 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d9c1c1c-859d-4b2a-9e6b-1a3d863df282" containerName="mariadb-database-create" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.992989 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="641f63ea-e837-4644-99af-c83045d72395" containerName="mariadb-account-create-update" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.992997 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4a29c27-3526-46bf-a3d2-3aacde054e78" containerName="mariadb-database-create" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.993004 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e6e56df-4807-42c7-a850-4369abb6e61e" containerName="mariadb-account-create-update" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.993017 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="db6357f7-cfa8-4f21-8b9f-223f828420c3" containerName="mariadb-database-create" Dec 01 17:43:03 crc kubenswrapper[4868]: I1201 17:43:03.993028 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="27bdb46a-6c28-4f18-82e7-50d25471969c" containerName="dnsmasq-dns" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.000060 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.027108 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-6lddf"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.054586 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-z6smf"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.055835 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.067745 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dxcx7" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.067974 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.068327 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.068435 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.068527 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.078909 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xwnx\" (UniqueName: \"kubernetes.io/projected/e6d1e182-191b-4828-8bba-aac485b02dc0-kube-api-access-4xwnx\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.079010 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.079033 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.079057 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-config\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.079085 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.079101 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.101004 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-z6smf"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.180186 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-config\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.180267 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-fernet-keys\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.180303 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-config-data\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.180324 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-scripts\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.180356 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.180375 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.180434 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvx42\" (UniqueName: \"kubernetes.io/projected/14bdadb2-2b54-41d1-a560-d68478f3bfb7-kube-api-access-rvx42\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.180483 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-combined-ca-bundle\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.180506 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xwnx\" (UniqueName: \"kubernetes.io/projected/e6d1e182-191b-4828-8bba-aac485b02dc0-kube-api-access-4xwnx\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.180568 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-credential-keys\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.180593 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.180618 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.181138 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-config\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.181464 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.181708 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.182323 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.182395 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.239890 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xwnx\" (UniqueName: \"kubernetes.io/projected/e6d1e182-191b-4828-8bba-aac485b02dc0-kube-api-access-4xwnx\") pod \"dnsmasq-dns-bbf5cc879-6lddf\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.282365 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-credential-keys\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.282422 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-fernet-keys\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.282442 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-config-data\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.282489 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-scripts\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.282565 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvx42\" (UniqueName: \"kubernetes.io/projected/14bdadb2-2b54-41d1-a560-d68478f3bfb7-kube-api-access-rvx42\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.282617 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-combined-ca-bundle\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.283845 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-68d44bd6f7-s9nkw"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.292867 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.297504 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.299729 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-config-data\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.297838 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.297898 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.313312 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-credential-keys\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.314563 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-combined-ca-bundle\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.329421 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.330891 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-fernet-keys\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.331267 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-scripts\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.342013 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-cc4w9" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.366629 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvx42\" (UniqueName: \"kubernetes.io/projected/14bdadb2-2b54-41d1-a560-d68478f3bfb7-kube-api-access-rvx42\") pod \"keystone-bootstrap-z6smf\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.402074 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-logs\") pod \"horizon-68d44bd6f7-s9nkw\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.402128 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l57k\" (UniqueName: \"kubernetes.io/projected/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-kube-api-access-4l57k\") pod \"horizon-68d44bd6f7-s9nkw\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.402186 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-horizon-secret-key\") pod \"horizon-68d44bd6f7-s9nkw\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.402207 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-scripts\") pod \"horizon-68d44bd6f7-s9nkw\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.402246 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-config-data\") pod \"horizon-68d44bd6f7-s9nkw\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.419736 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.421634 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68d44bd6f7-s9nkw"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.443691 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-fbgdl"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.472992 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fbgdl" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.483167 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.486766 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.488182 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-hhjr8" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.504086 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/27e82e73-8d20-4b1f-9514-11aad65f4331-config\") pod \"neutron-db-sync-fbgdl\" (UID: \"27e82e73-8d20-4b1f-9514-11aad65f4331\") " pod="openstack/neutron-db-sync-fbgdl" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.504146 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-logs\") pod \"horizon-68d44bd6f7-s9nkw\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.504170 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l57k\" (UniqueName: \"kubernetes.io/projected/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-kube-api-access-4l57k\") pod \"horizon-68d44bd6f7-s9nkw\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.504214 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e82e73-8d20-4b1f-9514-11aad65f4331-combined-ca-bundle\") pod \"neutron-db-sync-fbgdl\" (UID: \"27e82e73-8d20-4b1f-9514-11aad65f4331\") " pod="openstack/neutron-db-sync-fbgdl" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.504237 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s4fx\" (UniqueName: \"kubernetes.io/projected/27e82e73-8d20-4b1f-9514-11aad65f4331-kube-api-access-4s4fx\") pod \"neutron-db-sync-fbgdl\" (UID: \"27e82e73-8d20-4b1f-9514-11aad65f4331\") " pod="openstack/neutron-db-sync-fbgdl" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.504258 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-horizon-secret-key\") pod \"horizon-68d44bd6f7-s9nkw\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.504279 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-scripts\") pod \"horizon-68d44bd6f7-s9nkw\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.504318 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-config-data\") pod \"horizon-68d44bd6f7-s9nkw\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.505501 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-config-data\") pod \"horizon-68d44bd6f7-s9nkw\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.505734 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-logs\") pod \"horizon-68d44bd6f7-s9nkw\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.509697 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-fbgdl"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.510679 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-scripts\") pod \"horizon-68d44bd6f7-s9nkw\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.530397 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-hf8rq"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.533445 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hf8rq" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.536563 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-hf8rq"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.543073 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.543408 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kdp95" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.553036 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l57k\" (UniqueName: \"kubernetes.io/projected/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-kube-api-access-4l57k\") pod \"horizon-68d44bd6f7-s9nkw\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.583737 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-horizon-secret-key\") pod \"horizon-68d44bd6f7-s9nkw\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.606810 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.608326 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/27e82e73-8d20-4b1f-9514-11aad65f4331-config\") pod \"neutron-db-sync-fbgdl\" (UID: \"27e82e73-8d20-4b1f-9514-11aad65f4331\") " pod="openstack/neutron-db-sync-fbgdl" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.608427 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be92b184-744a-4da9-be7f-2441b7d1452d-combined-ca-bundle\") pod \"barbican-db-sync-hf8rq\" (UID: \"be92b184-744a-4da9-be7f-2441b7d1452d\") " pod="openstack/barbican-db-sync-hf8rq" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.608460 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8q54\" (UniqueName: \"kubernetes.io/projected/be92b184-744a-4da9-be7f-2441b7d1452d-kube-api-access-q8q54\") pod \"barbican-db-sync-hf8rq\" (UID: \"be92b184-744a-4da9-be7f-2441b7d1452d\") " pod="openstack/barbican-db-sync-hf8rq" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.608481 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e82e73-8d20-4b1f-9514-11aad65f4331-combined-ca-bundle\") pod \"neutron-db-sync-fbgdl\" (UID: \"27e82e73-8d20-4b1f-9514-11aad65f4331\") " pod="openstack/neutron-db-sync-fbgdl" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.608505 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s4fx\" (UniqueName: \"kubernetes.io/projected/27e82e73-8d20-4b1f-9514-11aad65f4331-kube-api-access-4s4fx\") pod \"neutron-db-sync-fbgdl\" (UID: \"27e82e73-8d20-4b1f-9514-11aad65f4331\") " pod="openstack/neutron-db-sync-fbgdl" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.608592 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/be92b184-744a-4da9-be7f-2441b7d1452d-db-sync-config-data\") pod \"barbican-db-sync-hf8rq\" (UID: \"be92b184-744a-4da9-be7f-2441b7d1452d\") " pod="openstack/barbican-db-sync-hf8rq" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.610383 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.625635 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.625850 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.636711 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e82e73-8d20-4b1f-9514-11aad65f4331-combined-ca-bundle\") pod \"neutron-db-sync-fbgdl\" (UID: \"27e82e73-8d20-4b1f-9514-11aad65f4331\") " pod="openstack/neutron-db-sync-fbgdl" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.636776 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.637340 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/27e82e73-8d20-4b1f-9514-11aad65f4331-config\") pod \"neutron-db-sync-fbgdl\" (UID: \"27e82e73-8d20-4b1f-9514-11aad65f4331\") " pod="openstack/neutron-db-sync-fbgdl" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.648467 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-vbxmg"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.649681 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.653648 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-78sw4" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.653933 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.654831 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s4fx\" (UniqueName: \"kubernetes.io/projected/27e82e73-8d20-4b1f-9514-11aad65f4331-kube-api-access-4s4fx\") pod \"neutron-db-sync-fbgdl\" (UID: \"27e82e73-8d20-4b1f-9514-11aad65f4331\") " pod="openstack/neutron-db-sync-fbgdl" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.660681 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.693417 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vbxmg"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.710948 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-config-data\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.711010 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-combined-ca-bundle\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.711038 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-scripts\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.711097 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/be92b184-744a-4da9-be7f-2441b7d1452d-db-sync-config-data\") pod \"barbican-db-sync-hf8rq\" (UID: \"be92b184-744a-4da9-be7f-2441b7d1452d\") " pod="openstack/barbican-db-sync-hf8rq" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.711128 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf9jh\" (UniqueName: \"kubernetes.io/projected/88c4f381-e8fa-474a-8f31-d0e2745ab83d-kube-api-access-mf9jh\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.711186 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbxtj\" (UniqueName: \"kubernetes.io/projected/c4416e29-710d-4683-9c40-92ca09b0e0bd-kube-api-access-lbxtj\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.711206 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.711226 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88c4f381-e8fa-474a-8f31-d0e2745ab83d-etc-machine-id\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.711244 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4416e29-710d-4683-9c40-92ca09b0e0bd-log-httpd\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.711261 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be92b184-744a-4da9-be7f-2441b7d1452d-combined-ca-bundle\") pod \"barbican-db-sync-hf8rq\" (UID: \"be92b184-744a-4da9-be7f-2441b7d1452d\") " pod="openstack/barbican-db-sync-hf8rq" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.711284 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4416e29-710d-4683-9c40-92ca09b0e0bd-run-httpd\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.711306 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-config-data\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.711331 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8q54\" (UniqueName: \"kubernetes.io/projected/be92b184-744a-4da9-be7f-2441b7d1452d-kube-api-access-q8q54\") pod \"barbican-db-sync-hf8rq\" (UID: \"be92b184-744a-4da9-be7f-2441b7d1452d\") " pod="openstack/barbican-db-sync-hf8rq" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.711364 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-scripts\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.711383 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.711398 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-db-sync-config-data\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.717564 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/be92b184-744a-4da9-be7f-2441b7d1452d-db-sync-config-data\") pod \"barbican-db-sync-hf8rq\" (UID: \"be92b184-744a-4da9-be7f-2441b7d1452d\") " pod="openstack/barbican-db-sync-hf8rq" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.718911 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be92b184-744a-4da9-be7f-2441b7d1452d-combined-ca-bundle\") pod \"barbican-db-sync-hf8rq\" (UID: \"be92b184-744a-4da9-be7f-2441b7d1452d\") " pod="openstack/barbican-db-sync-hf8rq" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.730117 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-mrzhs"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.731561 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.749421 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.749702 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.749858 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bsk5v" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.754242 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8q54\" (UniqueName: \"kubernetes.io/projected/be92b184-744a-4da9-be7f-2441b7d1452d-kube-api-access-q8q54\") pod \"barbican-db-sync-hf8rq\" (UID: \"be92b184-744a-4da9-be7f-2441b7d1452d\") " pod="openstack/barbican-db-sync-hf8rq" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.793607 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.817080 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-mrzhs"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818067 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-config-data\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818161 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-combined-ca-bundle\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818218 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-scripts\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818270 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf9jh\" (UniqueName: \"kubernetes.io/projected/88c4f381-e8fa-474a-8f31-d0e2745ab83d-kube-api-access-mf9jh\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818385 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-combined-ca-bundle\") pod \"placement-db-sync-mrzhs\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818460 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbxtj\" (UniqueName: \"kubernetes.io/projected/c4416e29-710d-4683-9c40-92ca09b0e0bd-kube-api-access-lbxtj\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818486 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwfcp\" (UniqueName: \"kubernetes.io/projected/f8a7a195-5f45-470b-acfc-db3d4b69ca12-kube-api-access-mwfcp\") pod \"placement-db-sync-mrzhs\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818530 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818585 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88c4f381-e8fa-474a-8f31-d0e2745ab83d-etc-machine-id\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818608 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4416e29-710d-4683-9c40-92ca09b0e0bd-log-httpd\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818714 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-scripts\") pod \"placement-db-sync-mrzhs\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818760 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4416e29-710d-4683-9c40-92ca09b0e0bd-run-httpd\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818787 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-config-data\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818845 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-config-data\") pod \"placement-db-sync-mrzhs\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818870 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8a7a195-5f45-470b-acfc-db3d4b69ca12-logs\") pod \"placement-db-sync-mrzhs\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.818929 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-scripts\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.819020 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-db-sync-config-data\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.819061 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.819187 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88c4f381-e8fa-474a-8f31-d0e2745ab83d-etc-machine-id\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.822651 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4416e29-710d-4683-9c40-92ca09b0e0bd-log-httpd\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.822706 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-config-data\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.822972 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4416e29-710d-4683-9c40-92ca09b0e0bd-run-httpd\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.831510 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-db-sync-config-data\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.832614 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-config-data\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.837429 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-scripts\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.838168 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.840612 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fbgdl" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.841374 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.846156 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-6lddf"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.856009 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf9jh\" (UniqueName: \"kubernetes.io/projected/88c4f381-e8fa-474a-8f31-d0e2745ab83d-kube-api-access-mf9jh\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.856817 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-scripts\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.860281 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-combined-ca-bundle\") pod \"cinder-db-sync-vbxmg\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.860881 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbxtj\" (UniqueName: \"kubernetes.io/projected/c4416e29-710d-4683-9c40-92ca09b0e0bd-kube-api-access-lbxtj\") pod \"ceilometer-0\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " pod="openstack/ceilometer-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.872991 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hf8rq" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.894340 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-68c7c6bb95-lm4km"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.896365 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.911391 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-v9rw6"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.913029 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.920722 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-config-data\") pod \"placement-db-sync-mrzhs\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.920762 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8a7a195-5f45-470b-acfc-db3d4b69ca12-logs\") pod \"placement-db-sync-mrzhs\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.920854 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-combined-ca-bundle\") pod \"placement-db-sync-mrzhs\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.920887 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwfcp\" (UniqueName: \"kubernetes.io/projected/f8a7a195-5f45-470b-acfc-db3d4b69ca12-kube-api-access-mwfcp\") pod \"placement-db-sync-mrzhs\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.920916 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-scripts\") pod \"placement-db-sync-mrzhs\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.923499 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8a7a195-5f45-470b-acfc-db3d4b69ca12-logs\") pod \"placement-db-sync-mrzhs\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.926965 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-config-data\") pod \"placement-db-sync-mrzhs\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.940512 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68c7c6bb95-lm4km"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.941191 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-combined-ca-bundle\") pod \"placement-db-sync-mrzhs\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.941338 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-scripts\") pod \"placement-db-sync-mrzhs\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.954776 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwfcp\" (UniqueName: \"kubernetes.io/projected/f8a7a195-5f45-470b-acfc-db3d4b69ca12-kube-api-access-mwfcp\") pod \"placement-db-sync-mrzhs\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.963803 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-v9rw6"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.980300 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.982154 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.984483 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.984690 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.984858 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5tv8n" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.986586 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 17:43:04 crc kubenswrapper[4868]: I1201 17:43:04.988790 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.000336 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.008121 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.012559 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.014355 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.017492 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.017735 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022109 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6cb38dd4-7c72-4b57-902f-67d10ccb033e-config-data\") pod \"horizon-68c7c6bb95-lm4km\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022159 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6cb38dd4-7c72-4b57-902f-67d10ccb033e-scripts\") pod \"horizon-68c7c6bb95-lm4km\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022177 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7265\" (UniqueName: \"kubernetes.io/projected/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-kube-api-access-b7265\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022198 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022229 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022251 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-config\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022276 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-scripts\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022296 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzp98\" (UniqueName: \"kubernetes.io/projected/904b45f4-66f7-4035-ae03-880709b69018-kube-api-access-pzp98\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022329 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l98nx\" (UniqueName: \"kubernetes.io/projected/6cb38dd4-7c72-4b57-902f-67d10ccb033e-kube-api-access-l98nx\") pod \"horizon-68c7c6bb95-lm4km\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022344 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022360 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-logs\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022686 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022709 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-config-data\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022739 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022764 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022777 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022810 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6cb38dd4-7c72-4b57-902f-67d10ccb033e-horizon-secret-key\") pod \"horizon-68c7c6bb95-lm4km\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022827 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cb38dd4-7c72-4b57-902f-67d10ccb033e-logs\") pod \"horizon-68c7c6bb95-lm4km\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.022841 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.023283 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.115785 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125058 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-config\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125129 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-scripts\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125175 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzp98\" (UniqueName: \"kubernetes.io/projected/904b45f4-66f7-4035-ae03-880709b69018-kube-api-access-pzp98\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125238 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l98nx\" (UniqueName: \"kubernetes.io/projected/6cb38dd4-7c72-4b57-902f-67d10ccb033e-kube-api-access-l98nx\") pod \"horizon-68c7c6bb95-lm4km\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125264 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125287 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-logs\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125363 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125394 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-config-data\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125420 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125459 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125478 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125505 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cb38dd4-7c72-4b57-902f-67d10ccb033e-logs\") pod \"horizon-68c7c6bb95-lm4km\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125525 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125557 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6cb38dd4-7c72-4b57-902f-67d10ccb033e-horizon-secret-key\") pod \"horizon-68c7c6bb95-lm4km\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125597 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6cb38dd4-7c72-4b57-902f-67d10ccb033e-config-data\") pod \"horizon-68c7c6bb95-lm4km\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125634 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6cb38dd4-7c72-4b57-902f-67d10ccb033e-scripts\") pod \"horizon-68c7c6bb95-lm4km\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125661 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7265\" (UniqueName: \"kubernetes.io/projected/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-kube-api-access-b7265\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125713 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.125745 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.126860 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.127701 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.129214 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.129510 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-logs\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.129718 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.132179 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6cb38dd4-7c72-4b57-902f-67d10ccb033e-config-data\") pod \"horizon-68c7c6bb95-lm4km\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.134421 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-config\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.135586 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.136223 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.137148 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cb38dd4-7c72-4b57-902f-67d10ccb033e-logs\") pod \"horizon-68c7c6bb95-lm4km\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.137984 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6cb38dd4-7c72-4b57-902f-67d10ccb033e-horizon-secret-key\") pod \"horizon-68c7c6bb95-lm4km\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.146613 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.146666 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6cb38dd4-7c72-4b57-902f-67d10ccb033e-scripts\") pod \"horizon-68c7c6bb95-lm4km\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.148761 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-6lddf"] Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.153422 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-scripts\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.161061 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-config-data\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.163435 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.170862 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l98nx\" (UniqueName: \"kubernetes.io/projected/6cb38dd4-7c72-4b57-902f-67d10ccb033e-kube-api-access-l98nx\") pod \"horizon-68c7c6bb95-lm4km\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.171540 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7265\" (UniqueName: \"kubernetes.io/projected/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-kube-api-access-b7265\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.191353 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzp98\" (UniqueName: \"kubernetes.io/projected/904b45f4-66f7-4035-ae03-880709b69018-kube-api-access-pzp98\") pod \"dnsmasq-dns-56df8fb6b7-v9rw6\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.227322 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.227376 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.227415 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.227445 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6dfb\" (UniqueName: \"kubernetes.io/projected/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-kube-api-access-q6dfb\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.227472 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.227508 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.227588 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-logs\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.227613 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: W1201 17:43:05.234049 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6d1e182_191b_4828_8bba_aac485b02dc0.slice/crio-26be17b78d9851babfee9b6b1aefef8cf5cf91cfec9d5aa3c980d1ebce4ccb18 WatchSource:0}: Error finding container 26be17b78d9851babfee9b6b1aefef8cf5cf91cfec9d5aa3c980d1ebce4ccb18: Status 404 returned error can't find the container with id 26be17b78d9851babfee9b6b1aefef8cf5cf91cfec9d5aa3c980d1ebce4ccb18 Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.234227 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.256318 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.301100 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-z6smf"] Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.328100 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.328600 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.328646 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.328686 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.328712 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6dfb\" (UniqueName: \"kubernetes.io/projected/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-kube-api-access-q6dfb\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.328770 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.328810 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.328844 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-logs\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.328866 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.333906 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.338702 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-logs\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.338981 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.353693 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.379398 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.394303 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.395681 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.407854 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6dfb\" (UniqueName: \"kubernetes.io/projected/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-kube-api-access-q6dfb\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.485195 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-z6smf" event={"ID":"14bdadb2-2b54-41d1-a560-d68478f3bfb7","Type":"ContainerStarted","Data":"ef12894ac3c40289bb726b578eb090822d080674b31237336e9afc04c641fe47"} Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.501443 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.503108 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" event={"ID":"e6d1e182-191b-4828-8bba-aac485b02dc0","Type":"ContainerStarted","Data":"26be17b78d9851babfee9b6b1aefef8cf5cf91cfec9d5aa3c980d1ebce4ccb18"} Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.620668 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.649755 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.802079 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-fbgdl"] Dec 01 17:43:05 crc kubenswrapper[4868]: I1201 17:43:05.945460 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-hf8rq"] Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.092756 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.105039 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68d44bd6f7-s9nkw"] Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.266339 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vbxmg"] Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.522165 4868 generic.go:334] "Generic (PLEG): container finished" podID="e6d1e182-191b-4828-8bba-aac485b02dc0" containerID="333b52ca1b636c0b43ebea3032a881e8a5db064d6f08d90cae92a1ee27fd6564" exitCode=0 Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.522271 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" event={"ID":"e6d1e182-191b-4828-8bba-aac485b02dc0","Type":"ContainerDied","Data":"333b52ca1b636c0b43ebea3032a881e8a5db064d6f08d90cae92a1ee27fd6564"} Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.525208 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hf8rq" event={"ID":"be92b184-744a-4da9-be7f-2441b7d1452d","Type":"ContainerStarted","Data":"35f89a0c283f556317e0f8e4c035881a4a3c239acbe138c27c3a2e89243693b4"} Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.538007 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vbxmg" event={"ID":"88c4f381-e8fa-474a-8f31-d0e2745ab83d","Type":"ContainerStarted","Data":"f1888bc29f4d53b5e5061a0f36554ee28f42ef2176bc89127498a7e34ef6f104"} Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.557421 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-z6smf" event={"ID":"14bdadb2-2b54-41d1-a560-d68478f3bfb7","Type":"ContainerStarted","Data":"345f997cae8a892e373c6b5bcfb2d6321751e959e76a526501d682d8bbac14fb"} Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.568530 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fbgdl" event={"ID":"27e82e73-8d20-4b1f-9514-11aad65f4331","Type":"ContainerStarted","Data":"58948cf462aa3d2b0ee2ea3d790a38977a1092272155fc8fcfaa5be21250a6d8"} Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.568577 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fbgdl" event={"ID":"27e82e73-8d20-4b1f-9514-11aad65f4331","Type":"ContainerStarted","Data":"a6b978b8ab98e7b955e528262abf9859c68d2fc6fd7d2a5713af259d84cbb89e"} Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.574115 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68d44bd6f7-s9nkw" event={"ID":"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a","Type":"ContainerStarted","Data":"fb604386519d190da339dd0a1f9c434c12a9b9a2222d95e98f94c669e3ceafc1"} Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.587480 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-z6smf" podStartSLOduration=3.587462345 podStartE2EDuration="3.587462345s" podCreationTimestamp="2025-12-01 17:43:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:06.579772511 +0000 UTC m=+1058.950882922" watchObservedRunningTime="2025-12-01 17:43:06.587462345 +0000 UTC m=+1058.958572756" Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.590147 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4416e29-710d-4683-9c40-92ca09b0e0bd","Type":"ContainerStarted","Data":"384485f0c86132095c26f2ac9c2f86bd5a951572999624cb434903ede448742e"} Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.613665 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-mrzhs"] Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.630380 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-fbgdl" podStartSLOduration=2.630358187 podStartE2EDuration="2.630358187s" podCreationTimestamp="2025-12-01 17:43:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:06.611554179 +0000 UTC m=+1058.982664610" watchObservedRunningTime="2025-12-01 17:43:06.630358187 +0000 UTC m=+1059.001468598" Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.666047 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-v9rw6"] Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.676653 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68c7c6bb95-lm4km"] Dec 01 17:43:06 crc kubenswrapper[4868]: W1201 17:43:06.681099 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6cb38dd4_7c72_4b57_902f_67d10ccb033e.slice/crio-0be8aabd81eaafc0e733abc84716510eba4a535c7945b2cb31e366111ccfaccc WatchSource:0}: Error finding container 0be8aabd81eaafc0e733abc84716510eba4a535c7945b2cb31e366111ccfaccc: Status 404 returned error can't find the container with id 0be8aabd81eaafc0e733abc84716510eba4a535c7945b2cb31e366111ccfaccc Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.818641 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 17:43:06 crc kubenswrapper[4868]: I1201 17:43:06.912303 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 17:43:06 crc kubenswrapper[4868]: W1201 17:43:06.976147 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb03a36ec_ecd5_4bf4_ad1e_99afed3ba83d.slice/crio-b3af8f9dd6eae405817a25e3f7e86eff4e41b5c54c889bae3d52ea75314d9412 WatchSource:0}: Error finding container b3af8f9dd6eae405817a25e3f7e86eff4e41b5c54c889bae3d52ea75314d9412: Status 404 returned error can't find the container with id b3af8f9dd6eae405817a25e3f7e86eff4e41b5c54c889bae3d52ea75314d9412 Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.026728 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.223962 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-ovsdbserver-nb\") pod \"e6d1e182-191b-4828-8bba-aac485b02dc0\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.224322 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-config\") pod \"e6d1e182-191b-4828-8bba-aac485b02dc0\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.224428 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-dns-swift-storage-0\") pod \"e6d1e182-191b-4828-8bba-aac485b02dc0\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.224466 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-ovsdbserver-sb\") pod \"e6d1e182-191b-4828-8bba-aac485b02dc0\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.224521 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xwnx\" (UniqueName: \"kubernetes.io/projected/e6d1e182-191b-4828-8bba-aac485b02dc0-kube-api-access-4xwnx\") pod \"e6d1e182-191b-4828-8bba-aac485b02dc0\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.224548 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-dns-svc\") pod \"e6d1e182-191b-4828-8bba-aac485b02dc0\" (UID: \"e6d1e182-191b-4828-8bba-aac485b02dc0\") " Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.248853 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6d1e182-191b-4828-8bba-aac485b02dc0-kube-api-access-4xwnx" (OuterVolumeSpecName: "kube-api-access-4xwnx") pod "e6d1e182-191b-4828-8bba-aac485b02dc0" (UID: "e6d1e182-191b-4828-8bba-aac485b02dc0"). InnerVolumeSpecName "kube-api-access-4xwnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.267219 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e6d1e182-191b-4828-8bba-aac485b02dc0" (UID: "e6d1e182-191b-4828-8bba-aac485b02dc0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.271739 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.287708 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-config" (OuterVolumeSpecName: "config") pod "e6d1e182-191b-4828-8bba-aac485b02dc0" (UID: "e6d1e182-191b-4828-8bba-aac485b02dc0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.326510 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e6d1e182-191b-4828-8bba-aac485b02dc0" (UID: "e6d1e182-191b-4828-8bba-aac485b02dc0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.328341 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.328557 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xwnx\" (UniqueName: \"kubernetes.io/projected/e6d1e182-191b-4828-8bba-aac485b02dc0-kube-api-access-4xwnx\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.328641 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.328722 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.339968 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68c7c6bb95-lm4km"] Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.359641 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e6d1e182-191b-4828-8bba-aac485b02dc0" (UID: "e6d1e182-191b-4828-8bba-aac485b02dc0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.394138 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e6d1e182-191b-4828-8bba-aac485b02dc0" (UID: "e6d1e182-191b-4828-8bba-aac485b02dc0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.418022 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.430436 4868 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.430473 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6d1e182-191b-4828-8bba-aac485b02dc0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.435484 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-84b8d76fb9-9nh6j"] Dec 01 17:43:07 crc kubenswrapper[4868]: E1201 17:43:07.435932 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6d1e182-191b-4828-8bba-aac485b02dc0" containerName="init" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.435959 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6d1e182-191b-4828-8bba-aac485b02dc0" containerName="init" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.436160 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6d1e182-191b-4828-8bba-aac485b02dc0" containerName="init" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.437261 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.449313 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-84b8d76fb9-9nh6j"] Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.546506 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.681672 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mrzhs" event={"ID":"f8a7a195-5f45-470b-acfc-db3d4b69ca12","Type":"ContainerStarted","Data":"dd94c5800e3af339816c8e271fd71e144a3c6e05232fac6934fb15684f063e6d"} Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.691477 4868 generic.go:334] "Generic (PLEG): container finished" podID="904b45f4-66f7-4035-ae03-880709b69018" containerID="d63c02da544826305cc2f69f33701eda03ecdce359ccd8793286eaa935851aee" exitCode=0 Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.691559 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" event={"ID":"904b45f4-66f7-4035-ae03-880709b69018","Type":"ContainerDied","Data":"d63c02da544826305cc2f69f33701eda03ecdce359ccd8793286eaa935851aee"} Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.691591 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" event={"ID":"904b45f4-66f7-4035-ae03-880709b69018","Type":"ContainerStarted","Data":"acbc9fcf12365522fa3744f1e56b60caf50c2ddff01bbba20c8e5523843c9cf6"} Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.725740 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-scripts\") pod \"horizon-84b8d76fb9-9nh6j\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.726808 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-config-data\") pod \"horizon-84b8d76fb9-9nh6j\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.726866 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zlbn\" (UniqueName: \"kubernetes.io/projected/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-kube-api-access-4zlbn\") pod \"horizon-84b8d76fb9-9nh6j\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.726931 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-horizon-secret-key\") pod \"horizon-84b8d76fb9-9nh6j\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.727022 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-logs\") pod \"horizon-84b8d76fb9-9nh6j\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.733338 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c7c6bb95-lm4km" event={"ID":"6cb38dd4-7c72-4b57-902f-67d10ccb033e","Type":"ContainerStarted","Data":"0be8aabd81eaafc0e733abc84716510eba4a535c7945b2cb31e366111ccfaccc"} Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.785283 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.785280 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-6lddf" event={"ID":"e6d1e182-191b-4828-8bba-aac485b02dc0","Type":"ContainerDied","Data":"26be17b78d9851babfee9b6b1aefef8cf5cf91cfec9d5aa3c980d1ebce4ccb18"} Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.785466 4868 scope.go:117] "RemoveContainer" containerID="333b52ca1b636c0b43ebea3032a881e8a5db064d6f08d90cae92a1ee27fd6564" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.791541 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e","Type":"ContainerStarted","Data":"be118bf089920fd11128a5368c477ed10505260e34c9e2476adc6c2dfc3628c4"} Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.814614 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d","Type":"ContainerStarted","Data":"b3af8f9dd6eae405817a25e3f7e86eff4e41b5c54c889bae3d52ea75314d9412"} Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.838858 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-horizon-secret-key\") pod \"horizon-84b8d76fb9-9nh6j\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.839023 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-logs\") pod \"horizon-84b8d76fb9-9nh6j\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.839443 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-scripts\") pod \"horizon-84b8d76fb9-9nh6j\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.850281 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-horizon-secret-key\") pod \"horizon-84b8d76fb9-9nh6j\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.851047 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-logs\") pod \"horizon-84b8d76fb9-9nh6j\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.851498 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-config-data\") pod \"horizon-84b8d76fb9-9nh6j\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.851624 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zlbn\" (UniqueName: \"kubernetes.io/projected/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-kube-api-access-4zlbn\") pod \"horizon-84b8d76fb9-9nh6j\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.853120 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-scripts\") pod \"horizon-84b8d76fb9-9nh6j\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.855379 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-config-data\") pod \"horizon-84b8d76fb9-9nh6j\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.881980 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zlbn\" (UniqueName: \"kubernetes.io/projected/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-kube-api-access-4zlbn\") pod \"horizon-84b8d76fb9-9nh6j\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.892023 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-6lddf"] Dec 01 17:43:07 crc kubenswrapper[4868]: I1201 17:43:07.911038 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-6lddf"] Dec 01 17:43:08 crc kubenswrapper[4868]: I1201 17:43:08.131815 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:08 crc kubenswrapper[4868]: I1201 17:43:08.230415 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6d1e182-191b-4828-8bba-aac485b02dc0" path="/var/lib/kubelet/pods/e6d1e182-191b-4828-8bba-aac485b02dc0/volumes" Dec 01 17:43:09 crc kubenswrapper[4868]: I1201 17:43:08.834642 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d","Type":"ContainerStarted","Data":"fa6e2508509774675ea0da54c013eabe42e4c8868e45311c25bb5f487790d76e"} Dec 01 17:43:09 crc kubenswrapper[4868]: I1201 17:43:08.839437 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" event={"ID":"904b45f4-66f7-4035-ae03-880709b69018","Type":"ContainerStarted","Data":"a8e492bc3fd84e965b5e5d33edba7fa6c37e3885c0432ef9fd726a8cd6f21d98"} Dec 01 17:43:09 crc kubenswrapper[4868]: I1201 17:43:08.839508 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:09 crc kubenswrapper[4868]: I1201 17:43:08.870713 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" podStartSLOduration=4.870692836 podStartE2EDuration="4.870692836s" podCreationTimestamp="2025-12-01 17:43:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:08.862737163 +0000 UTC m=+1061.233847594" watchObservedRunningTime="2025-12-01 17:43:08.870692836 +0000 UTC m=+1061.241803247" Dec 01 17:43:09 crc kubenswrapper[4868]: I1201 17:43:08.874523 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e","Type":"ContainerStarted","Data":"a516f0b147bfbcfdf8d9306810d0d899ff4e145a6d5019271197ebb9684d6bf8"} Dec 01 17:43:09 crc kubenswrapper[4868]: I1201 17:43:09.906770 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e","Type":"ContainerStarted","Data":"dfb8fa8048b3c2bd81aa56515745cfeb3c0afa45da27f3c07d209e28a6988b17"} Dec 01 17:43:09 crc kubenswrapper[4868]: I1201 17:43:09.907137 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" containerName="glance-log" containerID="cri-o://a516f0b147bfbcfdf8d9306810d0d899ff4e145a6d5019271197ebb9684d6bf8" gracePeriod=30 Dec 01 17:43:09 crc kubenswrapper[4868]: I1201 17:43:09.907537 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" containerName="glance-httpd" containerID="cri-o://dfb8fa8048b3c2bd81aa56515745cfeb3c0afa45da27f3c07d209e28a6988b17" gracePeriod=30 Dec 01 17:43:09 crc kubenswrapper[4868]: I1201 17:43:09.919771 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" containerName="glance-log" containerID="cri-o://fa6e2508509774675ea0da54c013eabe42e4c8868e45311c25bb5f487790d76e" gracePeriod=30 Dec 01 17:43:09 crc kubenswrapper[4868]: I1201 17:43:09.919904 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d","Type":"ContainerStarted","Data":"877352cd60f972ddcd1f4b4a64f6587d61bf4a94c451ac6643cd9b2be7e1fa78"} Dec 01 17:43:09 crc kubenswrapper[4868]: I1201 17:43:09.919990 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" containerName="glance-httpd" containerID="cri-o://877352cd60f972ddcd1f4b4a64f6587d61bf4a94c451ac6643cd9b2be7e1fa78" gracePeriod=30 Dec 01 17:43:09 crc kubenswrapper[4868]: I1201 17:43:09.958718 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.958698627 podStartE2EDuration="5.958698627s" podCreationTimestamp="2025-12-01 17:43:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:09.932216683 +0000 UTC m=+1062.303327094" watchObservedRunningTime="2025-12-01 17:43:09.958698627 +0000 UTC m=+1062.329809038" Dec 01 17:43:10 crc kubenswrapper[4868]: I1201 17:43:10.087745 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.087727988 podStartE2EDuration="6.087727988s" podCreationTimestamp="2025-12-01 17:43:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:09.957783894 +0000 UTC m=+1062.328894305" watchObservedRunningTime="2025-12-01 17:43:10.087727988 +0000 UTC m=+1062.458838399" Dec 01 17:43:10 crc kubenswrapper[4868]: I1201 17:43:10.101615 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-84b8d76fb9-9nh6j"] Dec 01 17:43:10 crc kubenswrapper[4868]: I1201 17:43:10.955670 4868 generic.go:334] "Generic (PLEG): container finished" podID="5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" containerID="dfb8fa8048b3c2bd81aa56515745cfeb3c0afa45da27f3c07d209e28a6988b17" exitCode=143 Dec 01 17:43:10 crc kubenswrapper[4868]: I1201 17:43:10.956048 4868 generic.go:334] "Generic (PLEG): container finished" podID="5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" containerID="a516f0b147bfbcfdf8d9306810d0d899ff4e145a6d5019271197ebb9684d6bf8" exitCode=143 Dec 01 17:43:10 crc kubenswrapper[4868]: I1201 17:43:10.955772 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e","Type":"ContainerDied","Data":"dfb8fa8048b3c2bd81aa56515745cfeb3c0afa45da27f3c07d209e28a6988b17"} Dec 01 17:43:10 crc kubenswrapper[4868]: I1201 17:43:10.956128 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e","Type":"ContainerDied","Data":"a516f0b147bfbcfdf8d9306810d0d899ff4e145a6d5019271197ebb9684d6bf8"} Dec 01 17:43:10 crc kubenswrapper[4868]: I1201 17:43:10.964977 4868 generic.go:334] "Generic (PLEG): container finished" podID="b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" containerID="877352cd60f972ddcd1f4b4a64f6587d61bf4a94c451ac6643cd9b2be7e1fa78" exitCode=143 Dec 01 17:43:10 crc kubenswrapper[4868]: I1201 17:43:10.965005 4868 generic.go:334] "Generic (PLEG): container finished" podID="b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" containerID="fa6e2508509774675ea0da54c013eabe42e4c8868e45311c25bb5f487790d76e" exitCode=143 Dec 01 17:43:10 crc kubenswrapper[4868]: I1201 17:43:10.965056 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d","Type":"ContainerDied","Data":"877352cd60f972ddcd1f4b4a64f6587d61bf4a94c451ac6643cd9b2be7e1fa78"} Dec 01 17:43:10 crc kubenswrapper[4868]: I1201 17:43:10.965084 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d","Type":"ContainerDied","Data":"fa6e2508509774675ea0da54c013eabe42e4c8868e45311c25bb5f487790d76e"} Dec 01 17:43:10 crc kubenswrapper[4868]: I1201 17:43:10.968312 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84b8d76fb9-9nh6j" event={"ID":"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20","Type":"ContainerStarted","Data":"b2e78c68534df4a916a3dee9d4b166e7c67a61e654f367911a6a6030e7db18b6"} Dec 01 17:43:12 crc kubenswrapper[4868]: I1201 17:43:12.991603 4868 generic.go:334] "Generic (PLEG): container finished" podID="14bdadb2-2b54-41d1-a560-d68478f3bfb7" containerID="345f997cae8a892e373c6b5bcfb2d6321751e959e76a526501d682d8bbac14fb" exitCode=0 Dec 01 17:43:12 crc kubenswrapper[4868]: I1201 17:43:12.992165 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-z6smf" event={"ID":"14bdadb2-2b54-41d1-a560-d68478f3bfb7","Type":"ContainerDied","Data":"345f997cae8a892e373c6b5bcfb2d6321751e959e76a526501d682d8bbac14fb"} Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.120351 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68d44bd6f7-s9nkw"] Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.195688 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-97868dbf6-djbqg"] Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.199421 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-97868dbf6-djbqg"] Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.199736 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.202295 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.247014 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-84b8d76fb9-9nh6j"] Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.295033 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb719ec-3812-4252-bdb9-1bc93aa35462-scripts\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.301009 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-horizon-secret-key\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.301196 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fdb719ec-3812-4252-bdb9-1bc93aa35462-logs\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.301296 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fdb719ec-3812-4252-bdb9-1bc93aa35462-config-data\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.301438 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-combined-ca-bundle\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.301466 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7kbw\" (UniqueName: \"kubernetes.io/projected/fdb719ec-3812-4252-bdb9-1bc93aa35462-kube-api-access-f7kbw\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.301505 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-horizon-tls-certs\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.344854 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-77567b6ccb-79skk"] Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.347783 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.353572 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77567b6ccb-79skk"] Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.403596 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-combined-ca-bundle\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.403638 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7kbw\" (UniqueName: \"kubernetes.io/projected/fdb719ec-3812-4252-bdb9-1bc93aa35462-kube-api-access-f7kbw\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.403661 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-horizon-tls-certs\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.403682 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb719ec-3812-4252-bdb9-1bc93aa35462-scripts\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.403725 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-horizon-secret-key\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.403775 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fdb719ec-3812-4252-bdb9-1bc93aa35462-logs\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.403812 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fdb719ec-3812-4252-bdb9-1bc93aa35462-config-data\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.405004 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fdb719ec-3812-4252-bdb9-1bc93aa35462-config-data\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.418746 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fdb719ec-3812-4252-bdb9-1bc93aa35462-logs\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.419253 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb719ec-3812-4252-bdb9-1bc93aa35462-scripts\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.419377 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-horizon-tls-certs\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.419523 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-combined-ca-bundle\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.421315 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-horizon-secret-key\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.424455 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7kbw\" (UniqueName: \"kubernetes.io/projected/fdb719ec-3812-4252-bdb9-1bc93aa35462-kube-api-access-f7kbw\") pod \"horizon-97868dbf6-djbqg\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.506256 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c50007b0-2158-4940-ac84-c60384469583-logs\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.506346 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c50007b0-2158-4940-ac84-c60384469583-scripts\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.506364 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c50007b0-2158-4940-ac84-c60384469583-config-data\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.506391 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-746zr\" (UniqueName: \"kubernetes.io/projected/c50007b0-2158-4940-ac84-c60384469583-kube-api-access-746zr\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.506414 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c50007b0-2158-4940-ac84-c60384469583-combined-ca-bundle\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.506452 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c50007b0-2158-4940-ac84-c60384469583-horizon-tls-certs\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.506475 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c50007b0-2158-4940-ac84-c60384469583-horizon-secret-key\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.531342 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.608077 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c50007b0-2158-4940-ac84-c60384469583-logs\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.608163 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c50007b0-2158-4940-ac84-c60384469583-scripts\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.608189 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c50007b0-2158-4940-ac84-c60384469583-config-data\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.608222 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-746zr\" (UniqueName: \"kubernetes.io/projected/c50007b0-2158-4940-ac84-c60384469583-kube-api-access-746zr\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.608249 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c50007b0-2158-4940-ac84-c60384469583-combined-ca-bundle\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.608293 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c50007b0-2158-4940-ac84-c60384469583-horizon-tls-certs\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.608317 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c50007b0-2158-4940-ac84-c60384469583-horizon-secret-key\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.609218 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c50007b0-2158-4940-ac84-c60384469583-logs\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.610102 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c50007b0-2158-4940-ac84-c60384469583-scripts\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.613554 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c50007b0-2158-4940-ac84-c60384469583-horizon-tls-certs\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.613680 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c50007b0-2158-4940-ac84-c60384469583-combined-ca-bundle\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.615554 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c50007b0-2158-4940-ac84-c60384469583-config-data\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.621284 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c50007b0-2158-4940-ac84-c60384469583-horizon-secret-key\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.629832 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-746zr\" (UniqueName: \"kubernetes.io/projected/c50007b0-2158-4940-ac84-c60384469583-kube-api-access-746zr\") pod \"horizon-77567b6ccb-79skk\" (UID: \"c50007b0-2158-4940-ac84-c60384469583\") " pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:14 crc kubenswrapper[4868]: I1201 17:43:14.686780 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:15 crc kubenswrapper[4868]: I1201 17:43:15.258169 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:15 crc kubenswrapper[4868]: I1201 17:43:15.400466 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-ksz2p"] Dec 01 17:43:15 crc kubenswrapper[4868]: I1201 17:43:15.401210 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" podUID="aefbe0d0-2918-4030-8ceb-7e686973efdb" containerName="dnsmasq-dns" containerID="cri-o://5abd2c337bc714a1a7939c3338e475adc671e7603ac1e9485f7bfd15b6f066ed" gracePeriod=10 Dec 01 17:43:16 crc kubenswrapper[4868]: I1201 17:43:16.031508 4868 generic.go:334] "Generic (PLEG): container finished" podID="aefbe0d0-2918-4030-8ceb-7e686973efdb" containerID="5abd2c337bc714a1a7939c3338e475adc671e7603ac1e9485f7bfd15b6f066ed" exitCode=0 Dec 01 17:43:16 crc kubenswrapper[4868]: I1201 17:43:16.031671 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" event={"ID":"aefbe0d0-2918-4030-8ceb-7e686973efdb","Type":"ContainerDied","Data":"5abd2c337bc714a1a7939c3338e475adc671e7603ac1e9485f7bfd15b6f066ed"} Dec 01 17:43:16 crc kubenswrapper[4868]: I1201 17:43:16.887048 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" podUID="aefbe0d0-2918-4030-8ceb-7e686973efdb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: connect: connection refused" Dec 01 17:43:18 crc kubenswrapper[4868]: I1201 17:43:18.081860 4868 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod903a0592-08cd-4d36-b634-87cadfb5c944"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod903a0592-08cd-4d36-b634-87cadfb5c944] : Timed out while waiting for systemd to remove kubepods-besteffort-pod903a0592_08cd_4d36_b634_87cadfb5c944.slice" Dec 01 17:43:18 crc kubenswrapper[4868]: E1201 17:43:18.082229 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod903a0592-08cd-4d36-b634-87cadfb5c944] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod903a0592-08cd-4d36-b634-87cadfb5c944] : Timed out while waiting for systemd to remove kubepods-besteffort-pod903a0592_08cd_4d36_b634_87cadfb5c944.slice" pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" podUID="903a0592-08cd-4d36-b634-87cadfb5c944" Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.057658 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-g2gmb" Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.085118 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-g2gmb"] Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.093111 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-g2gmb"] Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.237365 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.431305 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-config-data\") pod \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.431393 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvx42\" (UniqueName: \"kubernetes.io/projected/14bdadb2-2b54-41d1-a560-d68478f3bfb7-kube-api-access-rvx42\") pod \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.431538 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-credential-keys\") pod \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.433732 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-combined-ca-bundle\") pod \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.433761 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-scripts\") pod \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.433799 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-fernet-keys\") pod \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\" (UID: \"14bdadb2-2b54-41d1-a560-d68478f3bfb7\") " Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.440895 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14bdadb2-2b54-41d1-a560-d68478f3bfb7-kube-api-access-rvx42" (OuterVolumeSpecName: "kube-api-access-rvx42") pod "14bdadb2-2b54-41d1-a560-d68478f3bfb7" (UID: "14bdadb2-2b54-41d1-a560-d68478f3bfb7"). InnerVolumeSpecName "kube-api-access-rvx42". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.441660 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "14bdadb2-2b54-41d1-a560-d68478f3bfb7" (UID: "14bdadb2-2b54-41d1-a560-d68478f3bfb7"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.445215 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-scripts" (OuterVolumeSpecName: "scripts") pod "14bdadb2-2b54-41d1-a560-d68478f3bfb7" (UID: "14bdadb2-2b54-41d1-a560-d68478f3bfb7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.447711 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "14bdadb2-2b54-41d1-a560-d68478f3bfb7" (UID: "14bdadb2-2b54-41d1-a560-d68478f3bfb7"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.470892 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-config-data" (OuterVolumeSpecName: "config-data") pod "14bdadb2-2b54-41d1-a560-d68478f3bfb7" (UID: "14bdadb2-2b54-41d1-a560-d68478f3bfb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.476127 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14bdadb2-2b54-41d1-a560-d68478f3bfb7" (UID: "14bdadb2-2b54-41d1-a560-d68478f3bfb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.536725 4868 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.538532 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.538548 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.538559 4868 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.538568 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14bdadb2-2b54-41d1-a560-d68478f3bfb7-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:19 crc kubenswrapper[4868]: I1201 17:43:19.538577 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvx42\" (UniqueName: \"kubernetes.io/projected/14bdadb2-2b54-41d1-a560-d68478f3bfb7-kube-api-access-rvx42\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.075228 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-z6smf" event={"ID":"14bdadb2-2b54-41d1-a560-d68478f3bfb7","Type":"ContainerDied","Data":"ef12894ac3c40289bb726b578eb090822d080674b31237336e9afc04c641fe47"} Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.075278 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef12894ac3c40289bb726b578eb090822d080674b31237336e9afc04c641fe47" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.075586 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-z6smf" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.187210 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="903a0592-08cd-4d36-b634-87cadfb5c944" path="/var/lib/kubelet/pods/903a0592-08cd-4d36-b634-87cadfb5c944/volumes" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.316137 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-z6smf"] Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.324588 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-z6smf"] Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.485109 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-8stjq"] Dec 01 17:43:20 crc kubenswrapper[4868]: E1201 17:43:20.485494 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14bdadb2-2b54-41d1-a560-d68478f3bfb7" containerName="keystone-bootstrap" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.485505 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="14bdadb2-2b54-41d1-a560-d68478f3bfb7" containerName="keystone-bootstrap" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.485699 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="14bdadb2-2b54-41d1-a560-d68478f3bfb7" containerName="keystone-bootstrap" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.486296 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.493874 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8stjq"] Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.495274 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.495549 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.495723 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.495878 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dxcx7" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.496179 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.557334 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-combined-ca-bundle\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.557398 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-fernet-keys\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.557437 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fks8c\" (UniqueName: \"kubernetes.io/projected/ed070cb2-83dc-4f53-97ed-0ffff0baab82-kube-api-access-fks8c\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.557499 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-config-data\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.557565 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-credential-keys\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.557593 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-scripts\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.659270 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-credential-keys\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.659338 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-scripts\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.659370 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-combined-ca-bundle\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.659401 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-fernet-keys\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.659439 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fks8c\" (UniqueName: \"kubernetes.io/projected/ed070cb2-83dc-4f53-97ed-0ffff0baab82-kube-api-access-fks8c\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.659480 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-config-data\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.665554 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-scripts\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.665817 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-fernet-keys\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.665820 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-combined-ca-bundle\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.666414 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-config-data\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.672252 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-credential-keys\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.676449 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fks8c\" (UniqueName: \"kubernetes.io/projected/ed070cb2-83dc-4f53-97ed-0ffff0baab82-kube-api-access-fks8c\") pod \"keystone-bootstrap-8stjq\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.816249 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:20 crc kubenswrapper[4868]: I1201 17:43:20.904065 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.068226 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-internal-tls-certs\") pod \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.069351 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-combined-ca-bundle\") pod \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.069561 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-logs\") pod \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.069634 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6dfb\" (UniqueName: \"kubernetes.io/projected/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-kube-api-access-q6dfb\") pod \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.069850 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-config-data\") pod \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.070039 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-logs" (OuterVolumeSpecName: "logs") pod "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" (UID: "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.070199 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-httpd-run\") pod \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.070255 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-scripts\") pod \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.070282 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\" (UID: \"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d\") " Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.070487 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" (UID: "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.071132 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.071149 4868 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.075630 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" (UID: "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.075873 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-kube-api-access-q6dfb" (OuterVolumeSpecName: "kube-api-access-q6dfb") pod "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" (UID: "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d"). InnerVolumeSpecName "kube-api-access-q6dfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.077895 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-scripts" (OuterVolumeSpecName: "scripts") pod "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" (UID: "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.089833 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d","Type":"ContainerDied","Data":"b3af8f9dd6eae405817a25e3f7e86eff4e41b5c54c889bae3d52ea75314d9412"} Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.089899 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.089917 4868 scope.go:117] "RemoveContainer" containerID="877352cd60f972ddcd1f4b4a64f6587d61bf4a94c451ac6643cd9b2be7e1fa78" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.113999 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" (UID: "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.126152 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" (UID: "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.148093 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-config-data" (OuterVolumeSpecName: "config-data") pod "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" (UID: "b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.173434 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.173746 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6dfb\" (UniqueName: \"kubernetes.io/projected/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-kube-api-access-q6dfb\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.174205 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.174241 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.174289 4868 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.174308 4868 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.199397 4868 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.276990 4868 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.427126 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.440697 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.459240 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 17:43:21 crc kubenswrapper[4868]: E1201 17:43:21.459711 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" containerName="glance-httpd" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.459736 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" containerName="glance-httpd" Dec 01 17:43:21 crc kubenswrapper[4868]: E1201 17:43:21.459773 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" containerName="glance-log" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.459781 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" containerName="glance-log" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.460066 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" containerName="glance-httpd" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.460095 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" containerName="glance-log" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.462196 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.465118 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.465712 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.471057 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.585242 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.585293 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.585375 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/168bf194-0fdb-4919-89d9-cf91e0395c21-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.585414 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-config-data\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.585434 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/168bf194-0fdb-4919-89d9-cf91e0395c21-logs\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.585469 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjzv5\" (UniqueName: \"kubernetes.io/projected/168bf194-0fdb-4919-89d9-cf91e0395c21-kube-api-access-hjzv5\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.585494 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-scripts\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.585513 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.687130 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/168bf194-0fdb-4919-89d9-cf91e0395c21-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.687218 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-config-data\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.687247 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/168bf194-0fdb-4919-89d9-cf91e0395c21-logs\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.687296 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjzv5\" (UniqueName: \"kubernetes.io/projected/168bf194-0fdb-4919-89d9-cf91e0395c21-kube-api-access-hjzv5\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.687339 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-scripts\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.687365 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.687412 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.687437 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.687646 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.687656 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/168bf194-0fdb-4919-89d9-cf91e0395c21-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.688196 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/168bf194-0fdb-4919-89d9-cf91e0395c21-logs\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.697786 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-scripts\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.697827 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.698157 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.698872 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-config-data\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.709431 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjzv5\" (UniqueName: \"kubernetes.io/projected/168bf194-0fdb-4919-89d9-cf91e0395c21-kube-api-access-hjzv5\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.723207 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:43:21 crc kubenswrapper[4868]: I1201 17:43:21.804876 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.100304 4868 generic.go:334] "Generic (PLEG): container finished" podID="27e82e73-8d20-4b1f-9514-11aad65f4331" containerID="58948cf462aa3d2b0ee2ea3d790a38977a1092272155fc8fcfaa5be21250a6d8" exitCode=0 Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.100357 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fbgdl" event={"ID":"27e82e73-8d20-4b1f-9514-11aad65f4331","Type":"ContainerDied","Data":"58948cf462aa3d2b0ee2ea3d790a38977a1092272155fc8fcfaa5be21250a6d8"} Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.185073 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14bdadb2-2b54-41d1-a560-d68478f3bfb7" path="/var/lib/kubelet/pods/14bdadb2-2b54-41d1-a560-d68478f3bfb7/volumes" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.186154 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d" path="/var/lib/kubelet/pods/b03a36ec-ecd5-4bf4-ad1e-99afed3ba83d/volumes" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.615796 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.714003 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7265\" (UniqueName: \"kubernetes.io/projected/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-kube-api-access-b7265\") pod \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.714108 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-scripts\") pod \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.714145 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-combined-ca-bundle\") pod \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.714177 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.714224 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-logs\") pod \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.714313 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-config-data\") pod \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.714340 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-httpd-run\") pod \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.714549 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-public-tls-certs\") pod \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\" (UID: \"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e\") " Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.715274 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" (UID: "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.715352 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-logs" (OuterVolumeSpecName: "logs") pod "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" (UID: "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.721250 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-scripts" (OuterVolumeSpecName: "scripts") pod "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" (UID: "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.722043 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-kube-api-access-b7265" (OuterVolumeSpecName: "kube-api-access-b7265") pod "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" (UID: "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e"). InnerVolumeSpecName "kube-api-access-b7265". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.722246 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" (UID: "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.746958 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" (UID: "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.768823 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-config-data" (OuterVolumeSpecName: "config-data") pod "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" (UID: "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.771523 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" (UID: "5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.817112 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.817158 4868 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.817167 4868 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.817181 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7265\" (UniqueName: \"kubernetes.io/projected/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-kube-api-access-b7265\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.817192 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.817201 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.817240 4868 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.817252 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.832328 4868 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 01 17:43:22 crc kubenswrapper[4868]: I1201 17:43:22.919880 4868 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:23 crc kubenswrapper[4868]: E1201 17:43:23.063819 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 01 17:43:23 crc kubenswrapper[4868]: E1201 17:43:23.064060 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q8q54,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-hf8rq_openstack(be92b184-744a-4da9-be7f-2441b7d1452d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 17:43:23 crc kubenswrapper[4868]: E1201 17:43:23.065852 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-hf8rq" podUID="be92b184-744a-4da9-be7f-2441b7d1452d" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.114320 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e","Type":"ContainerDied","Data":"be118bf089920fd11128a5368c477ed10505260e34c9e2476adc6c2dfc3628c4"} Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.114537 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: E1201 17:43:23.120028 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-hf8rq" podUID="be92b184-744a-4da9-be7f-2441b7d1452d" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.168281 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.186663 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.201542 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 17:43:23 crc kubenswrapper[4868]: E1201 17:43:23.202199 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" containerName="glance-httpd" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.202221 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" containerName="glance-httpd" Dec 01 17:43:23 crc kubenswrapper[4868]: E1201 17:43:23.202265 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" containerName="glance-log" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.202275 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" containerName="glance-log" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.202520 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" containerName="glance-httpd" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.202546 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" containerName="glance-log" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.207556 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.212188 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.212625 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.212936 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.330613 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.330671 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-logs\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.330716 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqxxg\" (UniqueName: \"kubernetes.io/projected/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-kube-api-access-bqxxg\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.330781 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-scripts\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.330842 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.330903 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.330928 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-config-data\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.330988 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.433758 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.433875 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.433901 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-logs\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.433954 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqxxg\" (UniqueName: \"kubernetes.io/projected/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-kube-api-access-bqxxg\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.433995 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-scripts\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.434041 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.434094 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.434118 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-config-data\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.435233 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.435240 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.435529 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-logs\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.439812 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.440234 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.442689 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-config-data\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.444810 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-scripts\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.459256 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqxxg\" (UniqueName: \"kubernetes.io/projected/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-kube-api-access-bqxxg\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.467053 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " pod="openstack/glance-default-external-api-0" Dec 01 17:43:23 crc kubenswrapper[4868]: I1201 17:43:23.532734 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 17:43:24 crc kubenswrapper[4868]: I1201 17:43:24.184728 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e" path="/var/lib/kubelet/pods/5c950282-54bd-4fb4-a8fa-9c67f1fbfc2e/volumes" Dec 01 17:43:26 crc kubenswrapper[4868]: I1201 17:43:26.886928 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" podUID="aefbe0d0-2918-4030-8ceb-7e686973efdb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Dec 01 17:43:31 crc kubenswrapper[4868]: I1201 17:43:31.888107 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" podUID="aefbe0d0-2918-4030-8ceb-7e686973efdb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Dec 01 17:43:31 crc kubenswrapper[4868]: I1201 17:43:31.888851 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.150224 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.207427 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" event={"ID":"aefbe0d0-2918-4030-8ceb-7e686973efdb","Type":"ContainerDied","Data":"bfa4a88cff9f4c57b326525b5b7a9fe41c02e031260147985bff692754ba5c67"} Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.207474 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.246595 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-ovsdbserver-sb\") pod \"aefbe0d0-2918-4030-8ceb-7e686973efdb\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.246674 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-dns-swift-storage-0\") pod \"aefbe0d0-2918-4030-8ceb-7e686973efdb\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.246711 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-config\") pod \"aefbe0d0-2918-4030-8ceb-7e686973efdb\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.247152 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdlgm\" (UniqueName: \"kubernetes.io/projected/aefbe0d0-2918-4030-8ceb-7e686973efdb-kube-api-access-gdlgm\") pod \"aefbe0d0-2918-4030-8ceb-7e686973efdb\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.247212 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-dns-svc\") pod \"aefbe0d0-2918-4030-8ceb-7e686973efdb\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.247309 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-ovsdbserver-nb\") pod \"aefbe0d0-2918-4030-8ceb-7e686973efdb\" (UID: \"aefbe0d0-2918-4030-8ceb-7e686973efdb\") " Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.265920 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aefbe0d0-2918-4030-8ceb-7e686973efdb-kube-api-access-gdlgm" (OuterVolumeSpecName: "kube-api-access-gdlgm") pod "aefbe0d0-2918-4030-8ceb-7e686973efdb" (UID: "aefbe0d0-2918-4030-8ceb-7e686973efdb"). InnerVolumeSpecName "kube-api-access-gdlgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.296599 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aefbe0d0-2918-4030-8ceb-7e686973efdb" (UID: "aefbe0d0-2918-4030-8ceb-7e686973efdb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.297493 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aefbe0d0-2918-4030-8ceb-7e686973efdb" (UID: "aefbe0d0-2918-4030-8ceb-7e686973efdb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.301304 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aefbe0d0-2918-4030-8ceb-7e686973efdb" (UID: "aefbe0d0-2918-4030-8ceb-7e686973efdb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.301597 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-config" (OuterVolumeSpecName: "config") pod "aefbe0d0-2918-4030-8ceb-7e686973efdb" (UID: "aefbe0d0-2918-4030-8ceb-7e686973efdb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.314477 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aefbe0d0-2918-4030-8ceb-7e686973efdb" (UID: "aefbe0d0-2918-4030-8ceb-7e686973efdb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.349718 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.349749 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.349759 4868 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.349772 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.349781 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdlgm\" (UniqueName: \"kubernetes.io/projected/aefbe0d0-2918-4030-8ceb-7e686973efdb-kube-api-access-gdlgm\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.349791 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aefbe0d0-2918-4030-8ceb-7e686973efdb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:33 crc kubenswrapper[4868]: E1201 17:43:33.446826 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 01 17:43:33 crc kubenswrapper[4868]: E1201 17:43:33.447090 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nf5h679h66bh649h55fh657h5c9hc6h66fh87hcch546h75h657h96hd8h95h5b4hf5h5fchc7h676h59dh88h57h6hcbh97h55ch68fh584h59dq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lbxtj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(c4416e29-710d-4683-9c40-92ca09b0e0bd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.541426 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-ksz2p"] Dec 01 17:43:33 crc kubenswrapper[4868]: I1201 17:43:33.548243 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-ksz2p"] Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.185344 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aefbe0d0-2918-4030-8ceb-7e686973efdb" path="/var/lib/kubelet/pods/aefbe0d0-2918-4030-8ceb-7e686973efdb/volumes" Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.507721 4868 scope.go:117] "RemoveContainer" containerID="fa6e2508509774675ea0da54c013eabe42e4c8868e45311c25bb5f487790d76e" Dec 01 17:43:34 crc kubenswrapper[4868]: E1201 17:43:34.519671 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 01 17:43:34 crc kubenswrapper[4868]: E1201 17:43:34.519902 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mf9jh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-vbxmg_openstack(88c4f381-e8fa-474a-8f31-d0e2745ab83d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 17:43:34 crc kubenswrapper[4868]: E1201 17:43:34.522544 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-vbxmg" podUID="88c4f381-e8fa-474a-8f31-d0e2745ab83d" Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.686813 4868 scope.go:117] "RemoveContainer" containerID="dfb8fa8048b3c2bd81aa56515745cfeb3c0afa45da27f3c07d209e28a6988b17" Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.711903 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fbgdl" Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.772308 4868 scope.go:117] "RemoveContainer" containerID="a516f0b147bfbcfdf8d9306810d0d899ff4e145a6d5019271197ebb9684d6bf8" Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.782395 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e82e73-8d20-4b1f-9514-11aad65f4331-combined-ca-bundle\") pod \"27e82e73-8d20-4b1f-9514-11aad65f4331\" (UID: \"27e82e73-8d20-4b1f-9514-11aad65f4331\") " Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.782446 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s4fx\" (UniqueName: \"kubernetes.io/projected/27e82e73-8d20-4b1f-9514-11aad65f4331-kube-api-access-4s4fx\") pod \"27e82e73-8d20-4b1f-9514-11aad65f4331\" (UID: \"27e82e73-8d20-4b1f-9514-11aad65f4331\") " Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.783342 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/27e82e73-8d20-4b1f-9514-11aad65f4331-config\") pod \"27e82e73-8d20-4b1f-9514-11aad65f4331\" (UID: \"27e82e73-8d20-4b1f-9514-11aad65f4331\") " Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.815270 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27e82e73-8d20-4b1f-9514-11aad65f4331-kube-api-access-4s4fx" (OuterVolumeSpecName: "kube-api-access-4s4fx") pod "27e82e73-8d20-4b1f-9514-11aad65f4331" (UID: "27e82e73-8d20-4b1f-9514-11aad65f4331"). InnerVolumeSpecName "kube-api-access-4s4fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.850037 4868 scope.go:117] "RemoveContainer" containerID="5abd2c337bc714a1a7939c3338e475adc671e7603ac1e9485f7bfd15b6f066ed" Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.876242 4868 scope.go:117] "RemoveContainer" containerID="79dceeb275033ec3ee9c4e21e84ff0bf283589f4a030667c1cfb7c541c91e301" Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.886860 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4s4fx\" (UniqueName: \"kubernetes.io/projected/27e82e73-8d20-4b1f-9514-11aad65f4331-kube-api-access-4s4fx\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.926548 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e82e73-8d20-4b1f-9514-11aad65f4331-config" (OuterVolumeSpecName: "config") pod "27e82e73-8d20-4b1f-9514-11aad65f4331" (UID: "27e82e73-8d20-4b1f-9514-11aad65f4331"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.951008 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e82e73-8d20-4b1f-9514-11aad65f4331-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27e82e73-8d20-4b1f-9514-11aad65f4331" (UID: "27e82e73-8d20-4b1f-9514-11aad65f4331"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.968195 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77567b6ccb-79skk"] Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.989409 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e82e73-8d20-4b1f-9514-11aad65f4331-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:34 crc kubenswrapper[4868]: I1201 17:43:34.989439 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/27e82e73-8d20-4b1f-9514-11aad65f4331-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:34 crc kubenswrapper[4868]: W1201 17:43:34.993669 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc50007b0_2158_4940_ac84_c60384469583.slice/crio-c1721eaba11545a06ac3d76917014751d89efa24a89da0a307b91233dbf2964a WatchSource:0}: Error finding container c1721eaba11545a06ac3d76917014751d89efa24a89da0a307b91233dbf2964a: Status 404 returned error can't find the container with id c1721eaba11545a06ac3d76917014751d89efa24a89da0a307b91233dbf2964a Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.086241 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8stjq"] Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.114109 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-97868dbf6-djbqg"] Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.126097 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 17:43:35 crc kubenswrapper[4868]: W1201 17:43:35.141452 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdb719ec_3812_4252_bdb9_1bc93aa35462.slice/crio-795d41c1d61d507152f36a04c864649d2d3fb86f81c66835fb220bb1ba93df8d WatchSource:0}: Error finding container 795d41c1d61d507152f36a04c864649d2d3fb86f81c66835fb220bb1ba93df8d: Status 404 returned error can't find the container with id 795d41c1d61d507152f36a04c864649d2d3fb86f81c66835fb220bb1ba93df8d Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.172522 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.242028 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fbgdl" event={"ID":"27e82e73-8d20-4b1f-9514-11aad65f4331","Type":"ContainerDied","Data":"a6b978b8ab98e7b955e528262abf9859c68d2fc6fd7d2a5713af259d84cbb89e"} Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.242074 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6b978b8ab98e7b955e528262abf9859c68d2fc6fd7d2a5713af259d84cbb89e" Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.242177 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fbgdl" Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.263813 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68d44bd6f7-s9nkw" event={"ID":"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a","Type":"ContainerStarted","Data":"8aa42fb7693952e1198faa277e828f500f7f5c9df73038f7aed72d83081d0196"} Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.263898 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68d44bd6f7-s9nkw" event={"ID":"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a","Type":"ContainerStarted","Data":"3306b22dcff7cb03441971868a0d150d655932b2690fac92ca7e3d77390596f2"} Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.264056 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-68d44bd6f7-s9nkw" podUID="8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" containerName="horizon-log" containerID="cri-o://3306b22dcff7cb03441971868a0d150d655932b2690fac92ca7e3d77390596f2" gracePeriod=30 Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.264344 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-68d44bd6f7-s9nkw" podUID="8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" containerName="horizon" containerID="cri-o://8aa42fb7693952e1198faa277e828f500f7f5c9df73038f7aed72d83081d0196" gracePeriod=30 Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.269652 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c7c6bb95-lm4km" event={"ID":"6cb38dd4-7c72-4b57-902f-67d10ccb033e","Type":"ContainerStarted","Data":"88b6a6d8f215885bd8a47e36bdf1a0a076588c47a0f63a65c967e873a4be3b43"} Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.269704 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c7c6bb95-lm4km" event={"ID":"6cb38dd4-7c72-4b57-902f-67d10ccb033e","Type":"ContainerStarted","Data":"8ce65a2788466a9208f158286aa29b5856c5a7cf703453a0707de74b659b725b"} Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.271441 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-68c7c6bb95-lm4km" podUID="6cb38dd4-7c72-4b57-902f-67d10ccb033e" containerName="horizon-log" containerID="cri-o://8ce65a2788466a9208f158286aa29b5856c5a7cf703453a0707de74b659b725b" gracePeriod=30 Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.271585 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-68c7c6bb95-lm4km" podUID="6cb38dd4-7c72-4b57-902f-67d10ccb033e" containerName="horizon" containerID="cri-o://88b6a6d8f215885bd8a47e36bdf1a0a076588c47a0f63a65c967e873a4be3b43" gracePeriod=30 Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.280193 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-97868dbf6-djbqg" event={"ID":"fdb719ec-3812-4252-bdb9-1bc93aa35462","Type":"ContainerStarted","Data":"795d41c1d61d507152f36a04c864649d2d3fb86f81c66835fb220bb1ba93df8d"} Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.289911 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8stjq" event={"ID":"ed070cb2-83dc-4f53-97ed-0ffff0baab82","Type":"ContainerStarted","Data":"706762e63a46e0b4ded996e3caf3be3d57b4940dba54a53b7aae0ef5620222e6"} Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.298355 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-68d44bd6f7-s9nkw" podStartSLOduration=2.855144113 podStartE2EDuration="31.298331793s" podCreationTimestamp="2025-12-01 17:43:04 +0000 UTC" firstStartedPulling="2025-12-01 17:43:06.107584861 +0000 UTC m=+1058.478695272" lastFinishedPulling="2025-12-01 17:43:34.550772541 +0000 UTC m=+1086.921882952" observedRunningTime="2025-12-01 17:43:35.288954655 +0000 UTC m=+1087.660065066" watchObservedRunningTime="2025-12-01 17:43:35.298331793 +0000 UTC m=+1087.669442204" Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.304438 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"168bf194-0fdb-4919-89d9-cf91e0395c21","Type":"ContainerStarted","Data":"bfe6500751e8289183c54f560ff0195b530c0855a2dcef8214c466a35995ce58"} Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.318034 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-68c7c6bb95-lm4km" podStartSLOduration=3.5422205780000002 podStartE2EDuration="31.318011644s" podCreationTimestamp="2025-12-01 17:43:04 +0000 UTC" firstStartedPulling="2025-12-01 17:43:06.706651907 +0000 UTC m=+1059.077762318" lastFinishedPulling="2025-12-01 17:43:34.482442973 +0000 UTC m=+1086.853553384" observedRunningTime="2025-12-01 17:43:35.30799778 +0000 UTC m=+1087.679108211" watchObservedRunningTime="2025-12-01 17:43:35.318011644 +0000 UTC m=+1087.689122055" Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.320774 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mrzhs" event={"ID":"f8a7a195-5f45-470b-acfc-db3d4b69ca12","Type":"ContainerStarted","Data":"f779a6a1dfcf0682564edb209202e168825be2e615be4ed8dbcc17bc19bd18ca"} Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.328861 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77567b6ccb-79skk" event={"ID":"c50007b0-2158-4940-ac84-c60384469583","Type":"ContainerStarted","Data":"c1721eaba11545a06ac3d76917014751d89efa24a89da0a307b91233dbf2964a"} Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.331608 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-84b8d76fb9-9nh6j" podUID="09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" containerName="horizon-log" containerID="cri-o://b0b8ed13ae99e9296a2b0eadaacb695eae3ea49baffab2274a0b804636e28a2d" gracePeriod=30 Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.331731 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-84b8d76fb9-9nh6j" podUID="09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" containerName="horizon" containerID="cri-o://b8a80047e73bb1f55aaaf76795af5f147f5d7b15b8532f53e7b071523d9e34a4" gracePeriod=30 Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.332540 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84b8d76fb9-9nh6j" event={"ID":"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20","Type":"ContainerStarted","Data":"b8a80047e73bb1f55aaaf76795af5f147f5d7b15b8532f53e7b071523d9e34a4"} Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.332581 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84b8d76fb9-9nh6j" event={"ID":"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20","Type":"ContainerStarted","Data":"b0b8ed13ae99e9296a2b0eadaacb695eae3ea49baffab2274a0b804636e28a2d"} Dec 01 17:43:35 crc kubenswrapper[4868]: E1201 17:43:35.338826 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-vbxmg" podUID="88c4f381-e8fa-474a-8f31-d0e2745ab83d" Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.342126 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-mrzhs" podStartSLOduration=3.514708219 podStartE2EDuration="31.342102427s" podCreationTimestamp="2025-12-01 17:43:04 +0000 UTC" firstStartedPulling="2025-12-01 17:43:06.646460727 +0000 UTC m=+1059.017571138" lastFinishedPulling="2025-12-01 17:43:34.473854925 +0000 UTC m=+1086.844965346" observedRunningTime="2025-12-01 17:43:35.338490226 +0000 UTC m=+1087.709600637" watchObservedRunningTime="2025-12-01 17:43:35.342102427 +0000 UTC m=+1087.713212828" Dec 01 17:43:35 crc kubenswrapper[4868]: I1201 17:43:35.373665 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-84b8d76fb9-9nh6j" podStartSLOduration=4.042631265 podStartE2EDuration="28.373637309s" podCreationTimestamp="2025-12-01 17:43:07 +0000 UTC" firstStartedPulling="2025-12-01 17:43:10.211333932 +0000 UTC m=+1062.582444333" lastFinishedPulling="2025-12-01 17:43:34.542339966 +0000 UTC m=+1086.913450377" observedRunningTime="2025-12-01 17:43:35.357547109 +0000 UTC m=+1087.728657520" watchObservedRunningTime="2025-12-01 17:43:35.373637309 +0000 UTC m=+1087.744747730" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.053984 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-cftj7"] Dec 01 17:43:36 crc kubenswrapper[4868]: E1201 17:43:36.054984 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aefbe0d0-2918-4030-8ceb-7e686973efdb" containerName="init" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.055005 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="aefbe0d0-2918-4030-8ceb-7e686973efdb" containerName="init" Dec 01 17:43:36 crc kubenswrapper[4868]: E1201 17:43:36.055031 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e82e73-8d20-4b1f-9514-11aad65f4331" containerName="neutron-db-sync" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.055038 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e82e73-8d20-4b1f-9514-11aad65f4331" containerName="neutron-db-sync" Dec 01 17:43:36 crc kubenswrapper[4868]: E1201 17:43:36.055053 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aefbe0d0-2918-4030-8ceb-7e686973efdb" containerName="dnsmasq-dns" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.055060 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="aefbe0d0-2918-4030-8ceb-7e686973efdb" containerName="dnsmasq-dns" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.055244 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="27e82e73-8d20-4b1f-9514-11aad65f4331" containerName="neutron-db-sync" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.055262 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="aefbe0d0-2918-4030-8ceb-7e686973efdb" containerName="dnsmasq-dns" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.067978 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.102545 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-cftj7"] Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.126616 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.127068 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.127252 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-dns-svc\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.127391 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.127574 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrgdk\" (UniqueName: \"kubernetes.io/projected/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-kube-api-access-qrgdk\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.127681 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-config\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.163306 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-58c4c9c4cb-jdzxz"] Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.165048 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.174342 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-hhjr8" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.174685 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.174844 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.174934 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.230052 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn6tz\" (UniqueName: \"kubernetes.io/projected/c0cf4af0-a637-4d73-9f28-2ee6ea544584-kube-api-access-fn6tz\") pod \"neutron-58c4c9c4cb-jdzxz\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.230386 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-dns-svc\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.230413 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-config\") pod \"neutron-58c4c9c4cb-jdzxz\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.230433 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-httpd-config\") pod \"neutron-58c4c9c4cb-jdzxz\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.230453 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.230481 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrgdk\" (UniqueName: \"kubernetes.io/projected/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-kube-api-access-qrgdk\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.230505 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-config\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.230547 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.230569 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-combined-ca-bundle\") pod \"neutron-58c4c9c4cb-jdzxz\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.230627 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.230654 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-ovndb-tls-certs\") pod \"neutron-58c4c9c4cb-jdzxz\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.233020 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.235254 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-config\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.236015 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.240100 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-58c4c9c4cb-jdzxz"] Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.240138 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.252075 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.252617 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-dns-svc\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.283168 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrgdk\" (UniqueName: \"kubernetes.io/projected/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-kube-api-access-qrgdk\") pod \"dnsmasq-dns-6b7b667979-cftj7\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.333652 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn6tz\" (UniqueName: \"kubernetes.io/projected/c0cf4af0-a637-4d73-9f28-2ee6ea544584-kube-api-access-fn6tz\") pod \"neutron-58c4c9c4cb-jdzxz\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.333772 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-config\") pod \"neutron-58c4c9c4cb-jdzxz\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.333809 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-httpd-config\") pod \"neutron-58c4c9c4cb-jdzxz\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.333928 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-combined-ca-bundle\") pod \"neutron-58c4c9c4cb-jdzxz\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.334095 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-ovndb-tls-certs\") pod \"neutron-58c4c9c4cb-jdzxz\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.340812 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-ovndb-tls-certs\") pod \"neutron-58c4c9c4cb-jdzxz\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.347151 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-combined-ca-bundle\") pod \"neutron-58c4c9c4cb-jdzxz\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.351765 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-httpd-config\") pod \"neutron-58c4c9c4cb-jdzxz\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.352708 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-config\") pod \"neutron-58c4c9c4cb-jdzxz\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.367502 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn6tz\" (UniqueName: \"kubernetes.io/projected/c0cf4af0-a637-4d73-9f28-2ee6ea544584-kube-api-access-fn6tz\") pod \"neutron-58c4c9c4cb-jdzxz\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.413888 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.420844 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77567b6ccb-79skk" event={"ID":"c50007b0-2158-4940-ac84-c60384469583","Type":"ContainerStarted","Data":"32c7fca25eadeff828069145a94d0b67a9088f6e1d632ae561a3966637998ac7"} Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.420892 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77567b6ccb-79skk" event={"ID":"c50007b0-2158-4940-ac84-c60384469583","Type":"ContainerStarted","Data":"fa99fc830b149df594e58d4ec9304bb643db153b8c5f566e1c005a63e57ffd18"} Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.429034 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f6e2a84b-c570-44fa-9eba-1fbe11b761eb","Type":"ContainerStarted","Data":"1744ea545b02cbc388fac71ddd1ac84901c36a0c2140e2e14b3dc263e4cab5ca"} Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.446819 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4416e29-710d-4683-9c40-92ca09b0e0bd","Type":"ContainerStarted","Data":"e4775176a990584d72b3ad230e778ce5d004e058505e31837d7f84fd14e439f0"} Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.461663 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-77567b6ccb-79skk" podStartSLOduration=22.461636991 podStartE2EDuration="22.461636991s" podCreationTimestamp="2025-12-01 17:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:36.447241474 +0000 UTC m=+1088.818351895" watchObservedRunningTime="2025-12-01 17:43:36.461636991 +0000 UTC m=+1088.832747402" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.474525 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-97868dbf6-djbqg" event={"ID":"fdb719ec-3812-4252-bdb9-1bc93aa35462","Type":"ContainerStarted","Data":"5dbf91a124b9ea9b19e7613ba78db4552ecf9b204c056b4966d3fd691eb80b32"} Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.474586 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-97868dbf6-djbqg" event={"ID":"fdb719ec-3812-4252-bdb9-1bc93aa35462","Type":"ContainerStarted","Data":"98299ac37c9d398981045965e1ed7484fda45b7142252554239df2eab9ac0ba1"} Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.508182 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.509153 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8stjq" event={"ID":"ed070cb2-83dc-4f53-97ed-0ffff0baab82","Type":"ContainerStarted","Data":"1a5f10048b30cc2ad962b3f2b9c71894e13a17c8ca71dd957944f89f55c28409"} Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.527062 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-97868dbf6-djbqg" podStartSLOduration=22.527033154 podStartE2EDuration="22.527033154s" podCreationTimestamp="2025-12-01 17:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:36.518370443 +0000 UTC m=+1088.889480844" watchObservedRunningTime="2025-12-01 17:43:36.527033154 +0000 UTC m=+1088.898143565" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.889411 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-ksz2p" podUID="aefbe0d0-2918-4030-8ceb-7e686973efdb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.956142 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-8stjq" podStartSLOduration=16.956114546 podStartE2EDuration="16.956114546s" podCreationTimestamp="2025-12-01 17:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:36.567356249 +0000 UTC m=+1088.938466660" watchObservedRunningTime="2025-12-01 17:43:36.956114546 +0000 UTC m=+1089.327224957" Dec 01 17:43:36 crc kubenswrapper[4868]: I1201 17:43:36.970184 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-cftj7"] Dec 01 17:43:37 crc kubenswrapper[4868]: W1201 17:43:37.019048 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5da0876e_2895_4dd0_baf2_27ca9e5a1c30.slice/crio-3e4418add8137e6cce4260c3b0a33272d153d77806444ad109af022a0c578cc4 WatchSource:0}: Error finding container 3e4418add8137e6cce4260c3b0a33272d153d77806444ad109af022a0c578cc4: Status 404 returned error can't find the container with id 3e4418add8137e6cce4260c3b0a33272d153d77806444ad109af022a0c578cc4 Dec 01 17:43:37 crc kubenswrapper[4868]: I1201 17:43:37.221270 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-58c4c9c4cb-jdzxz"] Dec 01 17:43:37 crc kubenswrapper[4868]: I1201 17:43:37.576689 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58c4c9c4cb-jdzxz" event={"ID":"c0cf4af0-a637-4d73-9f28-2ee6ea544584","Type":"ContainerStarted","Data":"f9d94714ddd30af11be90141bd15ece7a3531d6c95627f1e4eeff3a1fb143d49"} Dec 01 17:43:37 crc kubenswrapper[4868]: I1201 17:43:37.598868 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"168bf194-0fdb-4919-89d9-cf91e0395c21","Type":"ContainerStarted","Data":"5dd285127895f6cb2cb5b9ca97f148b3d95e701f72c41175af3d807c0a595a59"} Dec 01 17:43:37 crc kubenswrapper[4868]: I1201 17:43:37.611663 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-cftj7" event={"ID":"5da0876e-2895-4dd0-baf2-27ca9e5a1c30","Type":"ContainerStarted","Data":"f54cc1ae6f27375662cb79010bca58c657a8ea75590a42b1332b7a595bcb5460"} Dec 01 17:43:37 crc kubenswrapper[4868]: I1201 17:43:37.611713 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-cftj7" event={"ID":"5da0876e-2895-4dd0-baf2-27ca9e5a1c30","Type":"ContainerStarted","Data":"3e4418add8137e6cce4260c3b0a33272d153d77806444ad109af022a0c578cc4"} Dec 01 17:43:37 crc kubenswrapper[4868]: I1201 17:43:37.619406 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hf8rq" event={"ID":"be92b184-744a-4da9-be7f-2441b7d1452d","Type":"ContainerStarted","Data":"9affd763fff7f399f265876d93260f2dc47a8b6ea25026ae7646d19a9d469498"} Dec 01 17:43:37 crc kubenswrapper[4868]: I1201 17:43:37.687146 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-hf8rq" podStartSLOduration=3.61541148 podStartE2EDuration="33.687125788s" podCreationTimestamp="2025-12-01 17:43:04 +0000 UTC" firstStartedPulling="2025-12-01 17:43:05.980025967 +0000 UTC m=+1058.351136378" lastFinishedPulling="2025-12-01 17:43:36.051740275 +0000 UTC m=+1088.422850686" observedRunningTime="2025-12-01 17:43:37.651863501 +0000 UTC m=+1090.022973932" watchObservedRunningTime="2025-12-01 17:43:37.687125788 +0000 UTC m=+1090.058236199" Dec 01 17:43:38 crc kubenswrapper[4868]: I1201 17:43:38.132448 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:43:38 crc kubenswrapper[4868]: I1201 17:43:38.642995 4868 generic.go:334] "Generic (PLEG): container finished" podID="f8a7a195-5f45-470b-acfc-db3d4b69ca12" containerID="f779a6a1dfcf0682564edb209202e168825be2e615be4ed8dbcc17bc19bd18ca" exitCode=0 Dec 01 17:43:38 crc kubenswrapper[4868]: I1201 17:43:38.643388 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mrzhs" event={"ID":"f8a7a195-5f45-470b-acfc-db3d4b69ca12","Type":"ContainerDied","Data":"f779a6a1dfcf0682564edb209202e168825be2e615be4ed8dbcc17bc19bd18ca"} Dec 01 17:43:38 crc kubenswrapper[4868]: I1201 17:43:38.652333 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58c4c9c4cb-jdzxz" event={"ID":"c0cf4af0-a637-4d73-9f28-2ee6ea544584","Type":"ContainerStarted","Data":"0886552474fe3abd4be9a720938c363213cb35a952612d81325fe20c40d49f01"} Dec 01 17:43:38 crc kubenswrapper[4868]: I1201 17:43:38.652383 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58c4c9c4cb-jdzxz" event={"ID":"c0cf4af0-a637-4d73-9f28-2ee6ea544584","Type":"ContainerStarted","Data":"acbef70eb8c395a23a94cccf92bb42860a4718e29265b97fa27fbbb999c88e83"} Dec 01 17:43:38 crc kubenswrapper[4868]: I1201 17:43:38.652677 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:43:38 crc kubenswrapper[4868]: I1201 17:43:38.655645 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f6e2a84b-c570-44fa-9eba-1fbe11b761eb","Type":"ContainerStarted","Data":"b9832a951b9d366bd17c83326764a4e46db1979d6f50477ef7fd5e2ea671ac9d"} Dec 01 17:43:38 crc kubenswrapper[4868]: I1201 17:43:38.701457 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"168bf194-0fdb-4919-89d9-cf91e0395c21","Type":"ContainerStarted","Data":"cf1b13312b06ed5a6f43881a3f655dcda1ff81353748e85d8a3e963732db1509"} Dec 01 17:43:38 crc kubenswrapper[4868]: I1201 17:43:38.734124 4868 generic.go:334] "Generic (PLEG): container finished" podID="5da0876e-2895-4dd0-baf2-27ca9e5a1c30" containerID="f54cc1ae6f27375662cb79010bca58c657a8ea75590a42b1332b7a595bcb5460" exitCode=0 Dec 01 17:43:38 crc kubenswrapper[4868]: I1201 17:43:38.734193 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-cftj7" event={"ID":"5da0876e-2895-4dd0-baf2-27ca9e5a1c30","Type":"ContainerDied","Data":"f54cc1ae6f27375662cb79010bca58c657a8ea75590a42b1332b7a595bcb5460"} Dec 01 17:43:38 crc kubenswrapper[4868]: I1201 17:43:38.734229 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-cftj7" event={"ID":"5da0876e-2895-4dd0-baf2-27ca9e5a1c30","Type":"ContainerStarted","Data":"e3bf0e44f1b4cfcae539a50f16335571e7dc7f254238644ee50304ecea1f324b"} Dec 01 17:43:38 crc kubenswrapper[4868]: I1201 17:43:38.734534 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:38 crc kubenswrapper[4868]: I1201 17:43:38.752362 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-58c4c9c4cb-jdzxz" podStartSLOduration=2.752334019 podStartE2EDuration="2.752334019s" podCreationTimestamp="2025-12-01 17:43:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:38.719812193 +0000 UTC m=+1091.090922594" watchObservedRunningTime="2025-12-01 17:43:38.752334019 +0000 UTC m=+1091.123444430" Dec 01 17:43:38 crc kubenswrapper[4868]: I1201 17:43:38.773063 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=17.773046757 podStartE2EDuration="17.773046757s" podCreationTimestamp="2025-12-01 17:43:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:38.770739178 +0000 UTC m=+1091.141849589" watchObservedRunningTime="2025-12-01 17:43:38.773046757 +0000 UTC m=+1091.144157168" Dec 01 17:43:38 crc kubenswrapper[4868]: I1201 17:43:38.824562 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-cftj7" podStartSLOduration=3.824540416 podStartE2EDuration="3.824540416s" podCreationTimestamp="2025-12-01 17:43:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:38.822068283 +0000 UTC m=+1091.193178704" watchObservedRunningTime="2025-12-01 17:43:38.824540416 +0000 UTC m=+1091.195650827" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.435901 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5c9f554f77-pjbhr"] Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.437834 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.441072 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.442204 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.457819 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c9f554f77-pjbhr"] Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.512854 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-internal-tls-certs\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.512913 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-ovndb-tls-certs\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.513009 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-httpd-config\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.513046 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-config\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.513131 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7shn\" (UniqueName: \"kubernetes.io/projected/cb459e0b-9096-4bc1-b553-41156deb1e39-kube-api-access-j7shn\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.513252 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-public-tls-certs\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.513299 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-combined-ca-bundle\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.615681 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7shn\" (UniqueName: \"kubernetes.io/projected/cb459e0b-9096-4bc1-b553-41156deb1e39-kube-api-access-j7shn\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.615800 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-public-tls-certs\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.615834 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-combined-ca-bundle\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.616028 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-internal-tls-certs\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.616053 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-ovndb-tls-certs\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.616099 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-httpd-config\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.616123 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-config\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.626860 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-public-tls-certs\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.665846 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-internal-tls-certs\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.666305 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-config\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.666510 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-httpd-config\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.666560 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-combined-ca-bundle\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.667077 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb459e0b-9096-4bc1-b553-41156deb1e39-ovndb-tls-certs\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.674556 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7shn\" (UniqueName: \"kubernetes.io/projected/cb459e0b-9096-4bc1-b553-41156deb1e39-kube-api-access-j7shn\") pod \"neutron-5c9f554f77-pjbhr\" (UID: \"cb459e0b-9096-4bc1-b553-41156deb1e39\") " pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.769916 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f6e2a84b-c570-44fa-9eba-1fbe11b761eb","Type":"ContainerStarted","Data":"277c8c1ea3ef347a4fac8ddbe7c0add43671c594e88b66a7eee55784ddd140eb"} Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.771043 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:39 crc kubenswrapper[4868]: I1201 17:43:39.812033 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=16.811931348999998 podStartE2EDuration="16.811931349s" podCreationTimestamp="2025-12-01 17:43:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:39.796521347 +0000 UTC m=+1092.167631748" watchObservedRunningTime="2025-12-01 17:43:39.811931349 +0000 UTC m=+1092.183041760" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.212433 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.234084 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-config-data\") pod \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.234181 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-scripts\") pod \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.234541 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8a7a195-5f45-470b-acfc-db3d4b69ca12-logs\") pod \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.234602 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwfcp\" (UniqueName: \"kubernetes.io/projected/f8a7a195-5f45-470b-acfc-db3d4b69ca12-kube-api-access-mwfcp\") pod \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.234668 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-combined-ca-bundle\") pod \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\" (UID: \"f8a7a195-5f45-470b-acfc-db3d4b69ca12\") " Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.235726 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8a7a195-5f45-470b-acfc-db3d4b69ca12-logs" (OuterVolumeSpecName: "logs") pod "f8a7a195-5f45-470b-acfc-db3d4b69ca12" (UID: "f8a7a195-5f45-470b-acfc-db3d4b69ca12"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.247064 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-scripts" (OuterVolumeSpecName: "scripts") pod "f8a7a195-5f45-470b-acfc-db3d4b69ca12" (UID: "f8a7a195-5f45-470b-acfc-db3d4b69ca12"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.264156 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8a7a195-5f45-470b-acfc-db3d4b69ca12-kube-api-access-mwfcp" (OuterVolumeSpecName: "kube-api-access-mwfcp") pod "f8a7a195-5f45-470b-acfc-db3d4b69ca12" (UID: "f8a7a195-5f45-470b-acfc-db3d4b69ca12"). InnerVolumeSpecName "kube-api-access-mwfcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.268064 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8a7a195-5f45-470b-acfc-db3d4b69ca12" (UID: "f8a7a195-5f45-470b-acfc-db3d4b69ca12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.274032 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-config-data" (OuterVolumeSpecName: "config-data") pod "f8a7a195-5f45-470b-acfc-db3d4b69ca12" (UID: "f8a7a195-5f45-470b-acfc-db3d4b69ca12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.341165 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.341203 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.341215 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8a7a195-5f45-470b-acfc-db3d4b69ca12-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.341225 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwfcp\" (UniqueName: \"kubernetes.io/projected/f8a7a195-5f45-470b-acfc-db3d4b69ca12-kube-api-access-mwfcp\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.341236 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a7a195-5f45-470b-acfc-db3d4b69ca12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.531547 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c9f554f77-pjbhr"] Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.802860 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c9f554f77-pjbhr" event={"ID":"cb459e0b-9096-4bc1-b553-41156deb1e39","Type":"ContainerStarted","Data":"8f5ca1e35ff368e141ee6c7211455bcb97f3136f66471a0f5cfd8b0f50810b60"} Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.809459 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mrzhs" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.813538 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mrzhs" event={"ID":"f8a7a195-5f45-470b-acfc-db3d4b69ca12","Type":"ContainerDied","Data":"dd94c5800e3af339816c8e271fd71e144a3c6e05232fac6934fb15684f063e6d"} Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.813607 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd94c5800e3af339816c8e271fd71e144a3c6e05232fac6934fb15684f063e6d" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.884434 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7d8466bd7b-sr5pl"] Dec 01 17:43:40 crc kubenswrapper[4868]: E1201 17:43:40.885027 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8a7a195-5f45-470b-acfc-db3d4b69ca12" containerName="placement-db-sync" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.885053 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8a7a195-5f45-470b-acfc-db3d4b69ca12" containerName="placement-db-sync" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.885336 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8a7a195-5f45-470b-acfc-db3d4b69ca12" containerName="placement-db-sync" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.895596 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.900201 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bsk5v" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.900489 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.900655 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.901056 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.903198 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 17:43:40 crc kubenswrapper[4868]: I1201 17:43:40.931000 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7d8466bd7b-sr5pl"] Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.055015 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f41cf3a5-eb10-42c0-93a1-f1195331f807-scripts\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.055058 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f41cf3a5-eb10-42c0-93a1-f1195331f807-config-data\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.055458 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f41cf3a5-eb10-42c0-93a1-f1195331f807-public-tls-certs\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.055597 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f41cf3a5-eb10-42c0-93a1-f1195331f807-logs\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.055636 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks4m2\" (UniqueName: \"kubernetes.io/projected/f41cf3a5-eb10-42c0-93a1-f1195331f807-kube-api-access-ks4m2\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.055722 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41cf3a5-eb10-42c0-93a1-f1195331f807-combined-ca-bundle\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.055804 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f41cf3a5-eb10-42c0-93a1-f1195331f807-internal-tls-certs\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.157967 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f41cf3a5-eb10-42c0-93a1-f1195331f807-public-tls-certs\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.158042 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f41cf3a5-eb10-42c0-93a1-f1195331f807-logs\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.158066 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks4m2\" (UniqueName: \"kubernetes.io/projected/f41cf3a5-eb10-42c0-93a1-f1195331f807-kube-api-access-ks4m2\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.158102 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41cf3a5-eb10-42c0-93a1-f1195331f807-combined-ca-bundle\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.158131 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f41cf3a5-eb10-42c0-93a1-f1195331f807-internal-tls-certs\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.158206 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f41cf3a5-eb10-42c0-93a1-f1195331f807-scripts\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.158225 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f41cf3a5-eb10-42c0-93a1-f1195331f807-config-data\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.158755 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f41cf3a5-eb10-42c0-93a1-f1195331f807-logs\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.169606 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f41cf3a5-eb10-42c0-93a1-f1195331f807-internal-tls-certs\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.169807 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41cf3a5-eb10-42c0-93a1-f1195331f807-combined-ca-bundle\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.170246 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f41cf3a5-eb10-42c0-93a1-f1195331f807-config-data\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.178222 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f41cf3a5-eb10-42c0-93a1-f1195331f807-scripts\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.183718 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f41cf3a5-eb10-42c0-93a1-f1195331f807-public-tls-certs\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.185607 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks4m2\" (UniqueName: \"kubernetes.io/projected/f41cf3a5-eb10-42c0-93a1-f1195331f807-kube-api-access-ks4m2\") pod \"placement-7d8466bd7b-sr5pl\" (UID: \"f41cf3a5-eb10-42c0-93a1-f1195331f807\") " pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.248669 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.805508 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.805850 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.837756 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c9f554f77-pjbhr" event={"ID":"cb459e0b-9096-4bc1-b553-41156deb1e39","Type":"ContainerStarted","Data":"699c0830a9ea64bb119a2ccac0ce3b7cf206e1e147f046767744d3e25560ae01"} Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.903474 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.904387 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:41 crc kubenswrapper[4868]: I1201 17:43:41.945092 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:42 crc kubenswrapper[4868]: I1201 17:43:42.015686 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7d8466bd7b-sr5pl"] Dec 01 17:43:42 crc kubenswrapper[4868]: I1201 17:43:42.864859 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c9f554f77-pjbhr" event={"ID":"cb459e0b-9096-4bc1-b553-41156deb1e39","Type":"ContainerStarted","Data":"6af2d367912dab33ace8609cea86ccf7f0240a55add63d483cbb29c80f5ef68c"} Dec 01 17:43:42 crc kubenswrapper[4868]: I1201 17:43:42.866354 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:43:42 crc kubenswrapper[4868]: I1201 17:43:42.878504 4868 generic.go:334] "Generic (PLEG): container finished" podID="be92b184-744a-4da9-be7f-2441b7d1452d" containerID="9affd763fff7f399f265876d93260f2dc47a8b6ea25026ae7646d19a9d469498" exitCode=0 Dec 01 17:43:42 crc kubenswrapper[4868]: I1201 17:43:42.878966 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hf8rq" event={"ID":"be92b184-744a-4da9-be7f-2441b7d1452d","Type":"ContainerDied","Data":"9affd763fff7f399f265876d93260f2dc47a8b6ea25026ae7646d19a9d469498"} Dec 01 17:43:42 crc kubenswrapper[4868]: I1201 17:43:42.895337 4868 generic.go:334] "Generic (PLEG): container finished" podID="ed070cb2-83dc-4f53-97ed-0ffff0baab82" containerID="1a5f10048b30cc2ad962b3f2b9c71894e13a17c8ca71dd957944f89f55c28409" exitCode=0 Dec 01 17:43:42 crc kubenswrapper[4868]: I1201 17:43:42.896982 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8stjq" event={"ID":"ed070cb2-83dc-4f53-97ed-0ffff0baab82","Type":"ContainerDied","Data":"1a5f10048b30cc2ad962b3f2b9c71894e13a17c8ca71dd957944f89f55c28409"} Dec 01 17:43:42 crc kubenswrapper[4868]: I1201 17:43:42.897262 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:42 crc kubenswrapper[4868]: I1201 17:43:42.899555 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5c9f554f77-pjbhr" podStartSLOduration=3.899530683 podStartE2EDuration="3.899530683s" podCreationTimestamp="2025-12-01 17:43:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:42.89147674 +0000 UTC m=+1095.262587151" watchObservedRunningTime="2025-12-01 17:43:42.899530683 +0000 UTC m=+1095.270641094" Dec 01 17:43:43 crc kubenswrapper[4868]: I1201 17:43:43.533647 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 17:43:43 crc kubenswrapper[4868]: I1201 17:43:43.533713 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 17:43:43 crc kubenswrapper[4868]: I1201 17:43:43.584825 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 17:43:43 crc kubenswrapper[4868]: I1201 17:43:43.592963 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 17:43:43 crc kubenswrapper[4868]: I1201 17:43:43.915351 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 17:43:43 crc kubenswrapper[4868]: I1201 17:43:43.915675 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 17:43:44 crc kubenswrapper[4868]: I1201 17:43:44.532277 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:44 crc kubenswrapper[4868]: I1201 17:43:44.532594 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:43:44 crc kubenswrapper[4868]: I1201 17:43:44.688754 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:44 crc kubenswrapper[4868]: I1201 17:43:44.688833 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:43:44 crc kubenswrapper[4868]: I1201 17:43:44.796313 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:43:44 crc kubenswrapper[4868]: I1201 17:43:44.923568 4868 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 17:43:45 crc kubenswrapper[4868]: I1201 17:43:45.234730 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:43:45 crc kubenswrapper[4868]: I1201 17:43:45.906826 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:45 crc kubenswrapper[4868]: I1201 17:43:45.944242 4868 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 17:43:45 crc kubenswrapper[4868]: I1201 17:43:45.944992 4868 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 17:43:46 crc kubenswrapper[4868]: I1201 17:43:46.416225 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:43:46 crc kubenswrapper[4868]: I1201 17:43:46.492306 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-v9rw6"] Dec 01 17:43:46 crc kubenswrapper[4868]: I1201 17:43:46.492804 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" podUID="904b45f4-66f7-4035-ae03-880709b69018" containerName="dnsmasq-dns" containerID="cri-o://a8e492bc3fd84e965b5e5d33edba7fa6c37e3885c0432ef9fd726a8cd6f21d98" gracePeriod=10 Dec 01 17:43:46 crc kubenswrapper[4868]: I1201 17:43:46.883961 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 17:43:46 crc kubenswrapper[4868]: I1201 17:43:46.963563 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 17:43:46 crc kubenswrapper[4868]: I1201 17:43:46.967652 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 17:43:46 crc kubenswrapper[4868]: I1201 17:43:46.968681 4868 generic.go:334] "Generic (PLEG): container finished" podID="904b45f4-66f7-4035-ae03-880709b69018" containerID="a8e492bc3fd84e965b5e5d33edba7fa6c37e3885c0432ef9fd726a8cd6f21d98" exitCode=0 Dec 01 17:43:46 crc kubenswrapper[4868]: I1201 17:43:46.969204 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" event={"ID":"904b45f4-66f7-4035-ae03-880709b69018","Type":"ContainerDied","Data":"a8e492bc3fd84e965b5e5d33edba7fa6c37e3885c0432ef9fd726a8cd6f21d98"} Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.653620 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.677683 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hf8rq" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.722049 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-combined-ca-bundle\") pod \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.722098 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/be92b184-744a-4da9-be7f-2441b7d1452d-db-sync-config-data\") pod \"be92b184-744a-4da9-be7f-2441b7d1452d\" (UID: \"be92b184-744a-4da9-be7f-2441b7d1452d\") " Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.722147 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-fernet-keys\") pod \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.722183 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-config-data\") pod \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.722360 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8q54\" (UniqueName: \"kubernetes.io/projected/be92b184-744a-4da9-be7f-2441b7d1452d-kube-api-access-q8q54\") pod \"be92b184-744a-4da9-be7f-2441b7d1452d\" (UID: \"be92b184-744a-4da9-be7f-2441b7d1452d\") " Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.722541 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be92b184-744a-4da9-be7f-2441b7d1452d-combined-ca-bundle\") pod \"be92b184-744a-4da9-be7f-2441b7d1452d\" (UID: \"be92b184-744a-4da9-be7f-2441b7d1452d\") " Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.722610 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-scripts\") pod \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.722651 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fks8c\" (UniqueName: \"kubernetes.io/projected/ed070cb2-83dc-4f53-97ed-0ffff0baab82-kube-api-access-fks8c\") pod \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.722685 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-credential-keys\") pod \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\" (UID: \"ed070cb2-83dc-4f53-97ed-0ffff0baab82\") " Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.728987 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ed070cb2-83dc-4f53-97ed-0ffff0baab82" (UID: "ed070cb2-83dc-4f53-97ed-0ffff0baab82"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.739781 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-scripts" (OuterVolumeSpecName: "scripts") pod "ed070cb2-83dc-4f53-97ed-0ffff0baab82" (UID: "ed070cb2-83dc-4f53-97ed-0ffff0baab82"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.743812 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be92b184-744a-4da9-be7f-2441b7d1452d-kube-api-access-q8q54" (OuterVolumeSpecName: "kube-api-access-q8q54") pod "be92b184-744a-4da9-be7f-2441b7d1452d" (UID: "be92b184-744a-4da9-be7f-2441b7d1452d"). InnerVolumeSpecName "kube-api-access-q8q54". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.746769 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be92b184-744a-4da9-be7f-2441b7d1452d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "be92b184-744a-4da9-be7f-2441b7d1452d" (UID: "be92b184-744a-4da9-be7f-2441b7d1452d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.749177 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ed070cb2-83dc-4f53-97ed-0ffff0baab82" (UID: "ed070cb2-83dc-4f53-97ed-0ffff0baab82"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.752199 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed070cb2-83dc-4f53-97ed-0ffff0baab82-kube-api-access-fks8c" (OuterVolumeSpecName: "kube-api-access-fks8c") pod "ed070cb2-83dc-4f53-97ed-0ffff0baab82" (UID: "ed070cb2-83dc-4f53-97ed-0ffff0baab82"). InnerVolumeSpecName "kube-api-access-fks8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.794309 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed070cb2-83dc-4f53-97ed-0ffff0baab82" (UID: "ed070cb2-83dc-4f53-97ed-0ffff0baab82"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.823834 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.823870 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fks8c\" (UniqueName: \"kubernetes.io/projected/ed070cb2-83dc-4f53-97ed-0ffff0baab82-kube-api-access-fks8c\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.823882 4868 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.823891 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.823901 4868 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/be92b184-744a-4da9-be7f-2441b7d1452d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.823909 4868 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.823918 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8q54\" (UniqueName: \"kubernetes.io/projected/be92b184-744a-4da9-be7f-2441b7d1452d-kube-api-access-q8q54\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.867979 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.869676 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be92b184-744a-4da9-be7f-2441b7d1452d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be92b184-744a-4da9-be7f-2441b7d1452d" (UID: "be92b184-744a-4da9-be7f-2441b7d1452d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:49 crc kubenswrapper[4868]: I1201 17:43:49.927261 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be92b184-744a-4da9-be7f-2441b7d1452d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.000993 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-config-data" (OuterVolumeSpecName: "config-data") pod "ed070cb2-83dc-4f53-97ed-0ffff0baab82" (UID: "ed070cb2-83dc-4f53-97ed-0ffff0baab82"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.029614 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzp98\" (UniqueName: \"kubernetes.io/projected/904b45f4-66f7-4035-ae03-880709b69018-kube-api-access-pzp98\") pod \"904b45f4-66f7-4035-ae03-880709b69018\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.029750 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-ovsdbserver-sb\") pod \"904b45f4-66f7-4035-ae03-880709b69018\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.029781 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-dns-svc\") pod \"904b45f4-66f7-4035-ae03-880709b69018\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.029862 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-config\") pod \"904b45f4-66f7-4035-ae03-880709b69018\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.029886 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-ovsdbserver-nb\") pod \"904b45f4-66f7-4035-ae03-880709b69018\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.029976 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-dns-swift-storage-0\") pod \"904b45f4-66f7-4035-ae03-880709b69018\" (UID: \"904b45f4-66f7-4035-ae03-880709b69018\") " Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.030397 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed070cb2-83dc-4f53-97ed-0ffff0baab82-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.037422 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/904b45f4-66f7-4035-ae03-880709b69018-kube-api-access-pzp98" (OuterVolumeSpecName: "kube-api-access-pzp98") pod "904b45f4-66f7-4035-ae03-880709b69018" (UID: "904b45f4-66f7-4035-ae03-880709b69018"). InnerVolumeSpecName "kube-api-access-pzp98". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.077351 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8stjq" event={"ID":"ed070cb2-83dc-4f53-97ed-0ffff0baab82","Type":"ContainerDied","Data":"706762e63a46e0b4ded996e3caf3be3d57b4940dba54a53b7aae0ef5620222e6"} Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.077781 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="706762e63a46e0b4ded996e3caf3be3d57b4940dba54a53b7aae0ef5620222e6" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.077882 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8stjq" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.095729 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d8466bd7b-sr5pl" event={"ID":"f41cf3a5-eb10-42c0-93a1-f1195331f807","Type":"ContainerStarted","Data":"57953ac79e2c11ff47c52a64f792578606296b5726fa57760129131eb04c2226"} Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.105263 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" event={"ID":"904b45f4-66f7-4035-ae03-880709b69018","Type":"ContainerDied","Data":"acbc9fcf12365522fa3744f1e56b60caf50c2ddff01bbba20c8e5523843c9cf6"} Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.105326 4868 scope.go:117] "RemoveContainer" containerID="a8e492bc3fd84e965b5e5d33edba7fa6c37e3885c0432ef9fd726a8cd6f21d98" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.105491 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-v9rw6" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.117988 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "904b45f4-66f7-4035-ae03-880709b69018" (UID: "904b45f4-66f7-4035-ae03-880709b69018"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.121449 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4416e29-710d-4683-9c40-92ca09b0e0bd","Type":"ContainerStarted","Data":"800e35a04f95ccb017cb1be7f474f630a8dcd89842c365e165b16af4786ea3a5"} Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.126418 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hf8rq" event={"ID":"be92b184-744a-4da9-be7f-2441b7d1452d","Type":"ContainerDied","Data":"35f89a0c283f556317e0f8e4c035881a4a3c239acbe138c27c3a2e89243693b4"} Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.126449 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35f89a0c283f556317e0f8e4c035881a4a3c239acbe138c27c3a2e89243693b4" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.126520 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hf8rq" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.131956 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "904b45f4-66f7-4035-ae03-880709b69018" (UID: "904b45f4-66f7-4035-ae03-880709b69018"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.131970 4868 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.132033 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzp98\" (UniqueName: \"kubernetes.io/projected/904b45f4-66f7-4035-ae03-880709b69018-kube-api-access-pzp98\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.135106 4868 scope.go:117] "RemoveContainer" containerID="d63c02da544826305cc2f69f33701eda03ecdce359ccd8793286eaa935851aee" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.144028 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "904b45f4-66f7-4035-ae03-880709b69018" (UID: "904b45f4-66f7-4035-ae03-880709b69018"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.148743 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-config" (OuterVolumeSpecName: "config") pod "904b45f4-66f7-4035-ae03-880709b69018" (UID: "904b45f4-66f7-4035-ae03-880709b69018"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.155672 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "904b45f4-66f7-4035-ae03-880709b69018" (UID: "904b45f4-66f7-4035-ae03-880709b69018"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.234305 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.234346 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.234358 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.234368 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/904b45f4-66f7-4035-ae03-880709b69018-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.473554 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-v9rw6"] Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.481274 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-v9rw6"] Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.803469 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5476cd6644-rgwc7"] Dec 01 17:43:50 crc kubenswrapper[4868]: E1201 17:43:50.803887 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed070cb2-83dc-4f53-97ed-0ffff0baab82" containerName="keystone-bootstrap" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.803903 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed070cb2-83dc-4f53-97ed-0ffff0baab82" containerName="keystone-bootstrap" Dec 01 17:43:50 crc kubenswrapper[4868]: E1201 17:43:50.803916 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be92b184-744a-4da9-be7f-2441b7d1452d" containerName="barbican-db-sync" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.803926 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="be92b184-744a-4da9-be7f-2441b7d1452d" containerName="barbican-db-sync" Dec 01 17:43:50 crc kubenswrapper[4868]: E1201 17:43:50.805899 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="904b45f4-66f7-4035-ae03-880709b69018" containerName="dnsmasq-dns" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.805920 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="904b45f4-66f7-4035-ae03-880709b69018" containerName="dnsmasq-dns" Dec 01 17:43:50 crc kubenswrapper[4868]: E1201 17:43:50.805933 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="904b45f4-66f7-4035-ae03-880709b69018" containerName="init" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.806007 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="904b45f4-66f7-4035-ae03-880709b69018" containerName="init" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.806205 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="904b45f4-66f7-4035-ae03-880709b69018" containerName="dnsmasq-dns" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.806221 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed070cb2-83dc-4f53-97ed-0ffff0baab82" containerName="keystone-bootstrap" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.806228 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="be92b184-744a-4da9-be7f-2441b7d1452d" containerName="barbican-db-sync" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.806848 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.811594 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.811732 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.815668 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dxcx7" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.815704 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.815825 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.819842 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.844378 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5476cd6644-rgwc7"] Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.949451 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-internal-tls-certs\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.949503 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-scripts\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.949563 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-credential-keys\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.949582 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-public-tls-certs\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.949602 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-config-data\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.949621 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-combined-ca-bundle\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.949642 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q46sq\" (UniqueName: \"kubernetes.io/projected/7366fab1-f630-4998-b189-d1b90a274235-kube-api-access-q46sq\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.949682 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-fernet-keys\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.980446 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-655b68fbf5-kxnvg"] Dec 01 17:43:50 crc kubenswrapper[4868]: I1201 17:43:50.994277 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.005457 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.005754 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.005891 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kdp95" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.009099 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-655b68fbf5-kxnvg"] Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.052771 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-credential-keys\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.052821 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-public-tls-certs\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.052847 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-config-data\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.052865 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-combined-ca-bundle\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.052886 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q46sq\" (UniqueName: \"kubernetes.io/projected/7366fab1-f630-4998-b189-d1b90a274235-kube-api-access-q46sq\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.052928 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-fernet-keys\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.053004 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-internal-tls-certs\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.053030 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-scripts\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.063001 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-79b8fddb46-xqzzs"] Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.064718 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.069665 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.073098 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-public-tls-certs\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.078275 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-combined-ca-bundle\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.078595 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-fernet-keys\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.085424 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-config-data\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.088562 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-internal-tls-certs\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.093869 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-79b8fddb46-xqzzs"] Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.099841 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-credential-keys\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.100139 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7366fab1-f630-4998-b189-d1b90a274235-scripts\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.117122 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q46sq\" (UniqueName: \"kubernetes.io/projected/7366fab1-f630-4998-b189-d1b90a274235-kube-api-access-q46sq\") pod \"keystone-5476cd6644-rgwc7\" (UID: \"7366fab1-f630-4998-b189-d1b90a274235\") " pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.117190 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-nmgvr"] Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.118999 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.138423 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.139113 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-nmgvr"] Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.233049 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d8466bd7b-sr5pl" event={"ID":"f41cf3a5-eb10-42c0-93a1-f1195331f807","Type":"ContainerStarted","Data":"5e052a250a87bfa53c3ab39660c825ec439664fb69f2910c509c4d0df686c181"} Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.233541 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d8466bd7b-sr5pl" event={"ID":"f41cf3a5-eb10-42c0-93a1-f1195331f807","Type":"ContainerStarted","Data":"43e5bfb9c1979c26820f51d5bc1a8536d7a4515a3f8b54af9fa88e4f133b3748"} Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.234254 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a4cab6a-d367-4eff-aab1-5b20b99f855e-logs\") pod \"barbican-keystone-listener-79b8fddb46-xqzzs\" (UID: \"0a4cab6a-d367-4eff-aab1-5b20b99f855e\") " pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.234368 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6fdfb743-319d-4a56-8182-442da79d3ed2-config-data-custom\") pod \"barbican-worker-655b68fbf5-kxnvg\" (UID: \"6fdfb743-319d-4a56-8182-442da79d3ed2\") " pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.234457 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.235059 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.244736 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.244868 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.245047 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.245194 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch8c5\" (UniqueName: \"kubernetes.io/projected/0a4cab6a-d367-4eff-aab1-5b20b99f855e-kube-api-access-ch8c5\") pod \"barbican-keystone-listener-79b8fddb46-xqzzs\" (UID: \"0a4cab6a-d367-4eff-aab1-5b20b99f855e\") " pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.245341 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fdfb743-319d-4a56-8182-442da79d3ed2-combined-ca-bundle\") pod \"barbican-worker-655b68fbf5-kxnvg\" (UID: \"6fdfb743-319d-4a56-8182-442da79d3ed2\") " pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.245440 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a4cab6a-d367-4eff-aab1-5b20b99f855e-config-data-custom\") pod \"barbican-keystone-listener-79b8fddb46-xqzzs\" (UID: \"0a4cab6a-d367-4eff-aab1-5b20b99f855e\") " pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.245510 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.245802 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fdfb743-319d-4a56-8182-442da79d3ed2-config-data\") pod \"barbican-worker-655b68fbf5-kxnvg\" (UID: \"6fdfb743-319d-4a56-8182-442da79d3ed2\") " pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.246457 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v46b9\" (UniqueName: \"kubernetes.io/projected/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-kube-api-access-v46b9\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.246567 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4cab6a-d367-4eff-aab1-5b20b99f855e-combined-ca-bundle\") pod \"barbican-keystone-listener-79b8fddb46-xqzzs\" (UID: \"0a4cab6a-d367-4eff-aab1-5b20b99f855e\") " pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.246753 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-config\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.246850 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fdfb743-319d-4a56-8182-442da79d3ed2-logs\") pod \"barbican-worker-655b68fbf5-kxnvg\" (UID: \"6fdfb743-319d-4a56-8182-442da79d3ed2\") " pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.247072 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm55m\" (UniqueName: \"kubernetes.io/projected/6fdfb743-319d-4a56-8182-442da79d3ed2-kube-api-access-hm55m\") pod \"barbican-worker-655b68fbf5-kxnvg\" (UID: \"6fdfb743-319d-4a56-8182-442da79d3ed2\") " pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.247331 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4cab6a-d367-4eff-aab1-5b20b99f855e-config-data\") pod \"barbican-keystone-listener-79b8fddb46-xqzzs\" (UID: \"0a4cab6a-d367-4eff-aab1-5b20b99f855e\") " pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.304349 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vbxmg" event={"ID":"88c4f381-e8fa-474a-8f31-d0e2745ab83d","Type":"ContainerStarted","Data":"517fb49a344e36a158dadd142d7cbf3c247e78907f8947945cade0df6d54b42b"} Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.331593 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6969f75cfd-5nchg"] Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.335143 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.340147 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.341786 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6969f75cfd-5nchg"] Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.347822 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7d8466bd7b-sr5pl" podStartSLOduration=11.347799954 podStartE2EDuration="11.347799954s" podCreationTimestamp="2025-12-01 17:43:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:51.321689458 +0000 UTC m=+1103.692799879" watchObservedRunningTime="2025-12-01 17:43:51.347799954 +0000 UTC m=+1103.718910365" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.361861 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-vbxmg" podStartSLOduration=4.174795799 podStartE2EDuration="47.361807988s" podCreationTimestamp="2025-12-01 17:43:04 +0000 UTC" firstStartedPulling="2025-12-01 17:43:06.278111308 +0000 UTC m=+1058.649221719" lastFinishedPulling="2025-12-01 17:43:49.465123507 +0000 UTC m=+1101.836233908" observedRunningTime="2025-12-01 17:43:51.353291844 +0000 UTC m=+1103.724402285" watchObservedRunningTime="2025-12-01 17:43:51.361807988 +0000 UTC m=+1103.732918419" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.369114 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.369447 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.369572 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch8c5\" (UniqueName: \"kubernetes.io/projected/0a4cab6a-d367-4eff-aab1-5b20b99f855e-kube-api-access-ch8c5\") pod \"barbican-keystone-listener-79b8fddb46-xqzzs\" (UID: \"0a4cab6a-d367-4eff-aab1-5b20b99f855e\") " pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.369662 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fdfb743-319d-4a56-8182-442da79d3ed2-combined-ca-bundle\") pod \"barbican-worker-655b68fbf5-kxnvg\" (UID: \"6fdfb743-319d-4a56-8182-442da79d3ed2\") " pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.369738 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a4cab6a-d367-4eff-aab1-5b20b99f855e-config-data-custom\") pod \"barbican-keystone-listener-79b8fddb46-xqzzs\" (UID: \"0a4cab6a-d367-4eff-aab1-5b20b99f855e\") " pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.369799 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.369880 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fdfb743-319d-4a56-8182-442da79d3ed2-config-data\") pod \"barbican-worker-655b68fbf5-kxnvg\" (UID: \"6fdfb743-319d-4a56-8182-442da79d3ed2\") " pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.370711 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx5wn\" (UniqueName: \"kubernetes.io/projected/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-kube-api-access-xx5wn\") pod \"barbican-api-6969f75cfd-5nchg\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.370809 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-combined-ca-bundle\") pod \"barbican-api-6969f75cfd-5nchg\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.370873 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-config-data-custom\") pod \"barbican-api-6969f75cfd-5nchg\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.370985 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v46b9\" (UniqueName: \"kubernetes.io/projected/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-kube-api-access-v46b9\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.371062 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4cab6a-d367-4eff-aab1-5b20b99f855e-combined-ca-bundle\") pod \"barbican-keystone-listener-79b8fddb46-xqzzs\" (UID: \"0a4cab6a-d367-4eff-aab1-5b20b99f855e\") " pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.371427 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-config\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.371517 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fdfb743-319d-4a56-8182-442da79d3ed2-logs\") pod \"barbican-worker-655b68fbf5-kxnvg\" (UID: \"6fdfb743-319d-4a56-8182-442da79d3ed2\") " pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.371590 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-logs\") pod \"barbican-api-6969f75cfd-5nchg\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.377254 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm55m\" (UniqueName: \"kubernetes.io/projected/6fdfb743-319d-4a56-8182-442da79d3ed2-kube-api-access-hm55m\") pod \"barbican-worker-655b68fbf5-kxnvg\" (UID: \"6fdfb743-319d-4a56-8182-442da79d3ed2\") " pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.379901 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.382821 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.388977 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-config\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.390759 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4cab6a-d367-4eff-aab1-5b20b99f855e-config-data\") pod \"barbican-keystone-listener-79b8fddb46-xqzzs\" (UID: \"0a4cab6a-d367-4eff-aab1-5b20b99f855e\") " pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.392310 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fdfb743-319d-4a56-8182-442da79d3ed2-logs\") pod \"barbican-worker-655b68fbf5-kxnvg\" (UID: \"6fdfb743-319d-4a56-8182-442da79d3ed2\") " pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.392622 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.392781 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-config-data\") pod \"barbican-api-6969f75cfd-5nchg\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.393077 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a4cab6a-d367-4eff-aab1-5b20b99f855e-logs\") pod \"barbican-keystone-listener-79b8fddb46-xqzzs\" (UID: \"0a4cab6a-d367-4eff-aab1-5b20b99f855e\") " pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.393493 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.393587 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6fdfb743-319d-4a56-8182-442da79d3ed2-config-data-custom\") pod \"barbican-worker-655b68fbf5-kxnvg\" (UID: \"6fdfb743-319d-4a56-8182-442da79d3ed2\") " pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.393433 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a4cab6a-d367-4eff-aab1-5b20b99f855e-logs\") pod \"barbican-keystone-listener-79b8fddb46-xqzzs\" (UID: \"0a4cab6a-d367-4eff-aab1-5b20b99f855e\") " pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.395550 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.396893 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fdfb743-319d-4a56-8182-442da79d3ed2-config-data\") pod \"barbican-worker-655b68fbf5-kxnvg\" (UID: \"6fdfb743-319d-4a56-8182-442da79d3ed2\") " pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.397911 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch8c5\" (UniqueName: \"kubernetes.io/projected/0a4cab6a-d367-4eff-aab1-5b20b99f855e-kube-api-access-ch8c5\") pod \"barbican-keystone-listener-79b8fddb46-xqzzs\" (UID: \"0a4cab6a-d367-4eff-aab1-5b20b99f855e\") " pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.403836 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6fdfb743-319d-4a56-8182-442da79d3ed2-config-data-custom\") pod \"barbican-worker-655b68fbf5-kxnvg\" (UID: \"6fdfb743-319d-4a56-8182-442da79d3ed2\") " pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.407997 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4cab6a-d367-4eff-aab1-5b20b99f855e-combined-ca-bundle\") pod \"barbican-keystone-listener-79b8fddb46-xqzzs\" (UID: \"0a4cab6a-d367-4eff-aab1-5b20b99f855e\") " pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.422000 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm55m\" (UniqueName: \"kubernetes.io/projected/6fdfb743-319d-4a56-8182-442da79d3ed2-kube-api-access-hm55m\") pod \"barbican-worker-655b68fbf5-kxnvg\" (UID: \"6fdfb743-319d-4a56-8182-442da79d3ed2\") " pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.423437 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v46b9\" (UniqueName: \"kubernetes.io/projected/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-kube-api-access-v46b9\") pod \"dnsmasq-dns-848cf88cfc-nmgvr\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.441817 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fdfb743-319d-4a56-8182-442da79d3ed2-combined-ca-bundle\") pod \"barbican-worker-655b68fbf5-kxnvg\" (UID: \"6fdfb743-319d-4a56-8182-442da79d3ed2\") " pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.444094 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4cab6a-d367-4eff-aab1-5b20b99f855e-config-data\") pod \"barbican-keystone-listener-79b8fddb46-xqzzs\" (UID: \"0a4cab6a-d367-4eff-aab1-5b20b99f855e\") " pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.467644 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a4cab6a-d367-4eff-aab1-5b20b99f855e-config-data-custom\") pod \"barbican-keystone-listener-79b8fddb46-xqzzs\" (UID: \"0a4cab6a-d367-4eff-aab1-5b20b99f855e\") " pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.496094 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx5wn\" (UniqueName: \"kubernetes.io/projected/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-kube-api-access-xx5wn\") pod \"barbican-api-6969f75cfd-5nchg\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.496151 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-combined-ca-bundle\") pod \"barbican-api-6969f75cfd-5nchg\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.496169 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-config-data-custom\") pod \"barbican-api-6969f75cfd-5nchg\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.496208 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-logs\") pod \"barbican-api-6969f75cfd-5nchg\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.496243 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-config-data\") pod \"barbican-api-6969f75cfd-5nchg\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.499641 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-logs\") pod \"barbican-api-6969f75cfd-5nchg\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.503579 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-combined-ca-bundle\") pod \"barbican-api-6969f75cfd-5nchg\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.511640 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-config-data-custom\") pod \"barbican-api-6969f75cfd-5nchg\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.517440 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-config-data\") pod \"barbican-api-6969f75cfd-5nchg\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.520407 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx5wn\" (UniqueName: \"kubernetes.io/projected/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-kube-api-access-xx5wn\") pod \"barbican-api-6969f75cfd-5nchg\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.573550 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.595276 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.612820 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.622814 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-655b68fbf5-kxnvg" Dec 01 17:43:51 crc kubenswrapper[4868]: I1201 17:43:51.807539 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5476cd6644-rgwc7"] Dec 01 17:43:51 crc kubenswrapper[4868]: W1201 17:43:51.826466 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7366fab1_f630_4998_b189_d1b90a274235.slice/crio-cf3fe62764f4d2ad3cd35edf63cb46cdc0f4ee4dbbe12be97eef225d766a7740 WatchSource:0}: Error finding container cf3fe62764f4d2ad3cd35edf63cb46cdc0f4ee4dbbe12be97eef225d766a7740: Status 404 returned error can't find the container with id cf3fe62764f4d2ad3cd35edf63cb46cdc0f4ee4dbbe12be97eef225d766a7740 Dec 01 17:43:52 crc kubenswrapper[4868]: I1201 17:43:52.190813 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="904b45f4-66f7-4035-ae03-880709b69018" path="/var/lib/kubelet/pods/904b45f4-66f7-4035-ae03-880709b69018/volumes" Dec 01 17:43:52 crc kubenswrapper[4868]: I1201 17:43:52.334077 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5476cd6644-rgwc7" event={"ID":"7366fab1-f630-4998-b189-d1b90a274235","Type":"ContainerStarted","Data":"cf3fe62764f4d2ad3cd35edf63cb46cdc0f4ee4dbbe12be97eef225d766a7740"} Dec 01 17:43:52 crc kubenswrapper[4868]: I1201 17:43:52.344395 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6969f75cfd-5nchg"] Dec 01 17:43:52 crc kubenswrapper[4868]: I1201 17:43:52.469061 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-79b8fddb46-xqzzs"] Dec 01 17:43:52 crc kubenswrapper[4868]: I1201 17:43:52.485758 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-nmgvr"] Dec 01 17:43:52 crc kubenswrapper[4868]: I1201 17:43:52.580562 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-655b68fbf5-kxnvg"] Dec 01 17:43:52 crc kubenswrapper[4868]: W1201 17:43:52.617079 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fdfb743_319d_4a56_8182_442da79d3ed2.slice/crio-14b0381532d09fbceebfd0ca6e026c74ef85fa56dcdaa6daaaccf37532e332cd WatchSource:0}: Error finding container 14b0381532d09fbceebfd0ca6e026c74ef85fa56dcdaa6daaaccf37532e332cd: Status 404 returned error can't find the container with id 14b0381532d09fbceebfd0ca6e026c74ef85fa56dcdaa6daaaccf37532e332cd Dec 01 17:43:53 crc kubenswrapper[4868]: I1201 17:43:53.357027 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" event={"ID":"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad","Type":"ContainerStarted","Data":"a5e60ed02c6f31d22c070cd7dd23a00fda084129e355d86bad9267d90ea43340"} Dec 01 17:43:53 crc kubenswrapper[4868]: I1201 17:43:53.365327 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" event={"ID":"0a4cab6a-d367-4eff-aab1-5b20b99f855e","Type":"ContainerStarted","Data":"25db726db818221be0f3b31b7c4e500ae035ca427bb6fcaf10769db9da384ba9"} Dec 01 17:43:53 crc kubenswrapper[4868]: I1201 17:43:53.369869 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6969f75cfd-5nchg" event={"ID":"c87b6efa-9ca1-4d35-9ff5-51e466b925d2","Type":"ContainerStarted","Data":"4b959695ec42dc9c5a1d7300b44624e3dd91cc3eae500f40687235a453d64ff7"} Dec 01 17:43:53 crc kubenswrapper[4868]: I1201 17:43:53.372138 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-655b68fbf5-kxnvg" event={"ID":"6fdfb743-319d-4a56-8182-442da79d3ed2","Type":"ContainerStarted","Data":"14b0381532d09fbceebfd0ca6e026c74ef85fa56dcdaa6daaaccf37532e332cd"} Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.460160 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6969f75cfd-5nchg" event={"ID":"c87b6efa-9ca1-4d35-9ff5-51e466b925d2","Type":"ContainerStarted","Data":"b733d416cd8a7287bce1bfd5b1b79fa10fe77f54f603d5c47f9d4ec32e984853"} Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.460681 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6969f75cfd-5nchg" event={"ID":"c87b6efa-9ca1-4d35-9ff5-51e466b925d2","Type":"ContainerStarted","Data":"ab3d83823b8b622d9d6fb5190ead32db337340e8c8e2e8cf3a7f154666001f13"} Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.462531 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.462630 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.480597 4868 generic.go:334] "Generic (PLEG): container finished" podID="b97acd99-8ef0-4c6d-a7d1-05f6c57450ad" containerID="704f8ef1fdfbea100f187f24fe378da5ea11adf7cf689d1e1b17cc8f6c1fe6cc" exitCode=0 Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.480996 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" event={"ID":"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad","Type":"ContainerDied","Data":"704f8ef1fdfbea100f187f24fe378da5ea11adf7cf689d1e1b17cc8f6c1fe6cc"} Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.488997 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6969f75cfd-5nchg" podStartSLOduration=3.488978411 podStartE2EDuration="3.488978411s" podCreationTimestamp="2025-12-01 17:43:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:54.482528144 +0000 UTC m=+1106.853638555" watchObservedRunningTime="2025-12-01 17:43:54.488978411 +0000 UTC m=+1106.860088822" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.491388 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5476cd6644-rgwc7" event={"ID":"7366fab1-f630-4998-b189-d1b90a274235","Type":"ContainerStarted","Data":"b1a219ae61c327cffde57dc2a8423b9b3590b97ea32eaec367b5c5e087fd1369"} Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.497246 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.535152 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-97868dbf6-djbqg" podUID="fdb719ec-3812-4252-bdb9-1bc93aa35462" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.578943 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5476cd6644-rgwc7" podStartSLOduration=4.578922446 podStartE2EDuration="4.578922446s" podCreationTimestamp="2025-12-01 17:43:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:54.564282935 +0000 UTC m=+1106.935393346" watchObservedRunningTime="2025-12-01 17:43:54.578922446 +0000 UTC m=+1106.950032857" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.694963 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-77567b6ccb-79skk" podUID="c50007b0-2158-4940-ac84-c60384469583" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.713256 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-77d99f96cd-fz4zl"] Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.714942 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.719238 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.719463 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.733456 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-77d99f96cd-fz4zl"] Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.807179 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/610e9c9c-cab0-4843-a120-5dde40363fd2-config-data-custom\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.807231 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/610e9c9c-cab0-4843-a120-5dde40363fd2-combined-ca-bundle\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.807293 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/610e9c9c-cab0-4843-a120-5dde40363fd2-config-data\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.807332 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/610e9c9c-cab0-4843-a120-5dde40363fd2-public-tls-certs\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.807478 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whzml\" (UniqueName: \"kubernetes.io/projected/610e9c9c-cab0-4843-a120-5dde40363fd2-kube-api-access-whzml\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.807528 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/610e9c9c-cab0-4843-a120-5dde40363fd2-logs\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.807665 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/610e9c9c-cab0-4843-a120-5dde40363fd2-internal-tls-certs\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.911613 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/610e9c9c-cab0-4843-a120-5dde40363fd2-config-data-custom\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.911669 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/610e9c9c-cab0-4843-a120-5dde40363fd2-combined-ca-bundle\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.911733 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/610e9c9c-cab0-4843-a120-5dde40363fd2-config-data\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.911778 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/610e9c9c-cab0-4843-a120-5dde40363fd2-public-tls-certs\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.911809 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whzml\" (UniqueName: \"kubernetes.io/projected/610e9c9c-cab0-4843-a120-5dde40363fd2-kube-api-access-whzml\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.911828 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/610e9c9c-cab0-4843-a120-5dde40363fd2-logs\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.911870 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/610e9c9c-cab0-4843-a120-5dde40363fd2-internal-tls-certs\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.912520 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/610e9c9c-cab0-4843-a120-5dde40363fd2-logs\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.917691 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/610e9c9c-cab0-4843-a120-5dde40363fd2-config-data-custom\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.917761 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/610e9c9c-cab0-4843-a120-5dde40363fd2-internal-tls-certs\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.920652 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/610e9c9c-cab0-4843-a120-5dde40363fd2-public-tls-certs\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.921071 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/610e9c9c-cab0-4843-a120-5dde40363fd2-config-data\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.929987 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whzml\" (UniqueName: \"kubernetes.io/projected/610e9c9c-cab0-4843-a120-5dde40363fd2-kube-api-access-whzml\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:54 crc kubenswrapper[4868]: I1201 17:43:54.938881 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/610e9c9c-cab0-4843-a120-5dde40363fd2-combined-ca-bundle\") pod \"barbican-api-77d99f96cd-fz4zl\" (UID: \"610e9c9c-cab0-4843-a120-5dde40363fd2\") " pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:55 crc kubenswrapper[4868]: I1201 17:43:55.094220 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:43:55 crc kubenswrapper[4868]: I1201 17:43:55.546048 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" event={"ID":"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad","Type":"ContainerStarted","Data":"fc643f8472870f6b7005fd6b073a21091bb9d65a0757a0e3b19dbb4ab4dee38d"} Dec 01 17:43:55 crc kubenswrapper[4868]: I1201 17:43:55.546542 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:43:55 crc kubenswrapper[4868]: I1201 17:43:55.573166 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" podStartSLOduration=4.573146114 podStartE2EDuration="4.573146114s" podCreationTimestamp="2025-12-01 17:43:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:43:55.570624975 +0000 UTC m=+1107.941735386" watchObservedRunningTime="2025-12-01 17:43:55.573146114 +0000 UTC m=+1107.944256525" Dec 01 17:43:55 crc kubenswrapper[4868]: I1201 17:43:55.680762 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-77d99f96cd-fz4zl"] Dec 01 17:43:57 crc kubenswrapper[4868]: I1201 17:43:57.591977 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77d99f96cd-fz4zl" event={"ID":"610e9c9c-cab0-4843-a120-5dde40363fd2","Type":"ContainerStarted","Data":"44e74f1726c77be12feab936d9589cfecc1aaffdc6adcbba0b500edc9cb4d6ca"} Dec 01 17:43:57 crc kubenswrapper[4868]: I1201 17:43:57.595664 4868 generic.go:334] "Generic (PLEG): container finished" podID="88c4f381-e8fa-474a-8f31-d0e2745ab83d" containerID="517fb49a344e36a158dadd142d7cbf3c247e78907f8947945cade0df6d54b42b" exitCode=0 Dec 01 17:43:57 crc kubenswrapper[4868]: I1201 17:43:57.595754 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vbxmg" event={"ID":"88c4f381-e8fa-474a-8f31-d0e2745ab83d","Type":"ContainerDied","Data":"517fb49a344e36a158dadd142d7cbf3c247e78907f8947945cade0df6d54b42b"} Dec 01 17:44:00 crc kubenswrapper[4868]: I1201 17:44:00.965568 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.160100 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-config-data\") pod \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.160168 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-combined-ca-bundle\") pod \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.160187 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-db-sync-config-data\") pod \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.160285 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88c4f381-e8fa-474a-8f31-d0e2745ab83d-etc-machine-id\") pod \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.160301 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-scripts\") pod \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.160348 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf9jh\" (UniqueName: \"kubernetes.io/projected/88c4f381-e8fa-474a-8f31-d0e2745ab83d-kube-api-access-mf9jh\") pod \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\" (UID: \"88c4f381-e8fa-474a-8f31-d0e2745ab83d\") " Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.160987 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88c4f381-e8fa-474a-8f31-d0e2745ab83d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "88c4f381-e8fa-474a-8f31-d0e2745ab83d" (UID: "88c4f381-e8fa-474a-8f31-d0e2745ab83d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.188188 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "88c4f381-e8fa-474a-8f31-d0e2745ab83d" (UID: "88c4f381-e8fa-474a-8f31-d0e2745ab83d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.190457 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-scripts" (OuterVolumeSpecName: "scripts") pod "88c4f381-e8fa-474a-8f31-d0e2745ab83d" (UID: "88c4f381-e8fa-474a-8f31-d0e2745ab83d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.191520 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88c4f381-e8fa-474a-8f31-d0e2745ab83d-kube-api-access-mf9jh" (OuterVolumeSpecName: "kube-api-access-mf9jh") pod "88c4f381-e8fa-474a-8f31-d0e2745ab83d" (UID: "88c4f381-e8fa-474a-8f31-d0e2745ab83d"). InnerVolumeSpecName "kube-api-access-mf9jh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.195666 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88c4f381-e8fa-474a-8f31-d0e2745ab83d" (UID: "88c4f381-e8fa-474a-8f31-d0e2745ab83d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.264265 4868 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88c4f381-e8fa-474a-8f31-d0e2745ab83d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.264311 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.264325 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf9jh\" (UniqueName: \"kubernetes.io/projected/88c4f381-e8fa-474a-8f31-d0e2745ab83d-kube-api-access-mf9jh\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.264337 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.264347 4868 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.297101 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-config-data" (OuterVolumeSpecName: "config-data") pod "88c4f381-e8fa-474a-8f31-d0e2745ab83d" (UID: "88c4f381-e8fa-474a-8f31-d0e2745ab83d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.367622 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88c4f381-e8fa-474a-8f31-d0e2745ab83d-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.615787 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.687860 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vbxmg" event={"ID":"88c4f381-e8fa-474a-8f31-d0e2745ab83d","Type":"ContainerDied","Data":"f1888bc29f4d53b5e5061a0f36554ee28f42ef2176bc89127498a7e34ef6f104"} Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.687907 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1888bc29f4d53b5e5061a0f36554ee28f42ef2176bc89127498a7e34ef6f104" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.688075 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vbxmg" Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.693925 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-cftj7"] Dec 01 17:44:01 crc kubenswrapper[4868]: I1201 17:44:01.694191 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-cftj7" podUID="5da0876e-2895-4dd0-baf2-27ca9e5a1c30" containerName="dnsmasq-dns" containerID="cri-o://e3bf0e44f1b4cfcae539a50f16335571e7dc7f254238644ee50304ecea1f324b" gracePeriod=10 Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.411244 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 17:44:02 crc kubenswrapper[4868]: E1201 17:44:02.411986 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88c4f381-e8fa-474a-8f31-d0e2745ab83d" containerName="cinder-db-sync" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.411997 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="88c4f381-e8fa-474a-8f31-d0e2745ab83d" containerName="cinder-db-sync" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.412261 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="88c4f381-e8fa-474a-8f31-d0e2745ab83d" containerName="cinder-db-sync" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.413344 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.424359 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.424639 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.424780 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.424900 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-78sw4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.430768 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-cfjj4"] Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.432509 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.464569 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.506643 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.506707 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-scripts\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.506733 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.506777 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b311f726-7ee0-452d-be0e-545b36e38666-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.506803 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-dns-svc\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.506866 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-config\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.506928 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.506974 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj6sb\" (UniqueName: \"kubernetes.io/projected/69c7a38b-c21b-46cd-a120-1d4cd1867941-kube-api-access-sj6sb\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.507049 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-config-data\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.507076 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.507188 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.507216 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pmqd\" (UniqueName: \"kubernetes.io/projected/b311f726-7ee0-452d-be0e-545b36e38666-kube-api-access-5pmqd\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.507704 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-cfjj4"] Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.579295 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.581540 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.586779 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.587557 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.615992 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.616055 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pmqd\" (UniqueName: \"kubernetes.io/projected/b311f726-7ee0-452d-be0e-545b36e38666-kube-api-access-5pmqd\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.616735 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.616761 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-scripts\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.618540 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.618601 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b311f726-7ee0-452d-be0e-545b36e38666-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.618625 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-dns-svc\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.618676 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-config\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.618700 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.618722 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj6sb\" (UniqueName: \"kubernetes.io/projected/69c7a38b-c21b-46cd-a120-1d4cd1867941-kube-api-access-sj6sb\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.618785 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-config-data\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.618811 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.619877 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-config\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.620047 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b311f726-7ee0-452d-be0e-545b36e38666-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.620602 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.620803 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.624358 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-dns-svc\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.629026 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.633760 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.634006 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-scripts\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.635988 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.644282 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-config-data\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.656692 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pmqd\" (UniqueName: \"kubernetes.io/projected/b311f726-7ee0-452d-be0e-545b36e38666-kube-api-access-5pmqd\") pod \"cinder-scheduler-0\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.674458 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj6sb\" (UniqueName: \"kubernetes.io/projected/69c7a38b-c21b-46cd-a120-1d4cd1867941-kube-api-access-sj6sb\") pod \"dnsmasq-dns-6578955fd5-cfjj4\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.725836 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.725915 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-config-data-custom\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.725987 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-scripts\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.726055 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/57856c20-00b1-4106-aa7e-a935af8c2814-etc-machine-id\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.726127 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57856c20-00b1-4106-aa7e-a935af8c2814-logs\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.726208 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-config-data\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.726280 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbdl2\" (UniqueName: \"kubernetes.io/projected/57856c20-00b1-4106-aa7e-a935af8c2814-kube-api-access-rbdl2\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.728887 4868 generic.go:334] "Generic (PLEG): container finished" podID="5da0876e-2895-4dd0-baf2-27ca9e5a1c30" containerID="e3bf0e44f1b4cfcae539a50f16335571e7dc7f254238644ee50304ecea1f324b" exitCode=0 Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.728998 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-cftj7" event={"ID":"5da0876e-2895-4dd0-baf2-27ca9e5a1c30","Type":"ContainerDied","Data":"e3bf0e44f1b4cfcae539a50f16335571e7dc7f254238644ee50304ecea1f324b"} Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.801541 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.816444 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.828863 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-config-data\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.828918 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbdl2\" (UniqueName: \"kubernetes.io/projected/57856c20-00b1-4106-aa7e-a935af8c2814-kube-api-access-rbdl2\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.829080 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.829104 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-config-data-custom\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.829130 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-scripts\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.829165 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/57856c20-00b1-4106-aa7e-a935af8c2814-etc-machine-id\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.829200 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57856c20-00b1-4106-aa7e-a935af8c2814-logs\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.829757 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57856c20-00b1-4106-aa7e-a935af8c2814-logs\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.832826 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/57856c20-00b1-4106-aa7e-a935af8c2814-etc-machine-id\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.836066 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-config-data\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.836237 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-scripts\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.839139 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.843757 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-config-data-custom\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.855441 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbdl2\" (UniqueName: \"kubernetes.io/projected/57856c20-00b1-4106-aa7e-a935af8c2814-kube-api-access-rbdl2\") pod \"cinder-api-0\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " pod="openstack/cinder-api-0" Dec 01 17:44:02 crc kubenswrapper[4868]: I1201 17:44:02.900121 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.285225 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.398602 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-config\") pod \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.398680 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-dns-svc\") pod \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.398761 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrgdk\" (UniqueName: \"kubernetes.io/projected/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-kube-api-access-qrgdk\") pod \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.398784 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-dns-swift-storage-0\") pod \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.398811 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-ovsdbserver-nb\") pod \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.398988 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-ovsdbserver-sb\") pod \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\" (UID: \"5da0876e-2895-4dd0-baf2-27ca9e5a1c30\") " Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.420378 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.422491 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-kube-api-access-qrgdk" (OuterVolumeSpecName: "kube-api-access-qrgdk") pod "5da0876e-2895-4dd0-baf2-27ca9e5a1c30" (UID: "5da0876e-2895-4dd0-baf2-27ca9e5a1c30"). InnerVolumeSpecName "kube-api-access-qrgdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.504054 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrgdk\" (UniqueName: \"kubernetes.io/projected/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-kube-api-access-qrgdk\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.531967 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-97868dbf6-djbqg" podUID="fdb719ec-3812-4252-bdb9-1bc93aa35462" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.626101 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-cfjj4"] Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.634352 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.709546 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5da0876e-2895-4dd0-baf2-27ca9e5a1c30" (UID: "5da0876e-2895-4dd0-baf2-27ca9e5a1c30"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.711985 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:04 crc kubenswrapper[4868]: E1201 17:44:04.780864 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="c4416e29-710d-4683-9c40-92ca09b0e0bd" Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.840539 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-config" (OuterVolumeSpecName: "config") pod "5da0876e-2895-4dd0-baf2-27ca9e5a1c30" (UID: "5da0876e-2895-4dd0-baf2-27ca9e5a1c30"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.894350 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4416e29-710d-4683-9c40-92ca09b0e0bd","Type":"ContainerStarted","Data":"7024addc687087043e4fb44941d29311179976a32ceaac0d20843ca1623961eb"} Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.894538 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerName="ceilometer-notification-agent" containerID="cri-o://e4775176a990584d72b3ad230e778ce5d004e058505e31837d7f84fd14e439f0" gracePeriod=30 Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.894787 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.895123 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerName="proxy-httpd" containerID="cri-o://7024addc687087043e4fb44941d29311179976a32ceaac0d20843ca1623961eb" gracePeriod=30 Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.895175 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerName="sg-core" containerID="cri-o://800e35a04f95ccb017cb1be7f474f630a8dcd89842c365e165b16af4786ea3a5" gracePeriod=30 Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.904920 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b311f726-7ee0-452d-be0e-545b36e38666","Type":"ContainerStarted","Data":"dcb11f7c381ef006b26266a9347e2fba80bbf60089e9abc217f6afe2b0233e09"} Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.916820 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.917357 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" event={"ID":"69c7a38b-c21b-46cd-a120-1d4cd1867941","Type":"ContainerStarted","Data":"f52cc5e86f2c22fc91faa0fbb6c3af111cf5af50496bbd2f78be42de9cf55a9f"} Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.927812 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"57856c20-00b1-4106-aa7e-a935af8c2814","Type":"ContainerStarted","Data":"d198b76742ac2e6df80489cd31fb920d60a48b75800ea7af18820a1247594585"} Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.935515 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-cftj7" event={"ID":"5da0876e-2895-4dd0-baf2-27ca9e5a1c30","Type":"ContainerDied","Data":"3e4418add8137e6cce4260c3b0a33272d153d77806444ad109af022a0c578cc4"} Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.935568 4868 scope.go:117] "RemoveContainer" containerID="e3bf0e44f1b4cfcae539a50f16335571e7dc7f254238644ee50304ecea1f324b" Dec 01 17:44:04 crc kubenswrapper[4868]: I1201 17:44:04.935705 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-cftj7" Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.130557 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5da0876e-2895-4dd0-baf2-27ca9e5a1c30" (UID: "5da0876e-2895-4dd0-baf2-27ca9e5a1c30"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.153853 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5da0876e-2895-4dd0-baf2-27ca9e5a1c30" (UID: "5da0876e-2895-4dd0-baf2-27ca9e5a1c30"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.178665 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5da0876e-2895-4dd0-baf2-27ca9e5a1c30" (UID: "5da0876e-2895-4dd0-baf2-27ca9e5a1c30"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.181018 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.227777 4868 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.227833 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.227844 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5da0876e-2895-4dd0-baf2-27ca9e5a1c30-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.352831 4868 scope.go:117] "RemoveContainer" containerID="f54cc1ae6f27375662cb79010bca58c657a8ea75590a42b1332b7a595bcb5460" Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.388865 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.733237 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-cftj7"] Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.753616 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-cftj7"] Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.761892 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.995749 4868 generic.go:334] "Generic (PLEG): container finished" podID="6cb38dd4-7c72-4b57-902f-67d10ccb033e" containerID="88b6a6d8f215885bd8a47e36bdf1a0a076588c47a0f63a65c967e873a4be3b43" exitCode=137 Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.996258 4868 generic.go:334] "Generic (PLEG): container finished" podID="6cb38dd4-7c72-4b57-902f-67d10ccb033e" containerID="8ce65a2788466a9208f158286aa29b5856c5a7cf703453a0707de74b659b725b" exitCode=137 Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.996316 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c7c6bb95-lm4km" event={"ID":"6cb38dd4-7c72-4b57-902f-67d10ccb033e","Type":"ContainerDied","Data":"88b6a6d8f215885bd8a47e36bdf1a0a076588c47a0f63a65c967e873a4be3b43"} Dec 01 17:44:05 crc kubenswrapper[4868]: I1201 17:44:05.996352 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c7c6bb95-lm4km" event={"ID":"6cb38dd4-7c72-4b57-902f-67d10ccb033e","Type":"ContainerDied","Data":"8ce65a2788466a9208f158286aa29b5856c5a7cf703453a0707de74b659b725b"} Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.006918 4868 generic.go:334] "Generic (PLEG): container finished" podID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerID="800e35a04f95ccb017cb1be7f474f630a8dcd89842c365e165b16af4786ea3a5" exitCode=2 Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.006976 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4416e29-710d-4683-9c40-92ca09b0e0bd","Type":"ContainerDied","Data":"800e35a04f95ccb017cb1be7f474f630a8dcd89842c365e165b16af4786ea3a5"} Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.009047 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-655b68fbf5-kxnvg" event={"ID":"6fdfb743-319d-4a56-8182-442da79d3ed2","Type":"ContainerStarted","Data":"99675c04cd08d2e40e5f464b685ebfd4f3effbb387e2f5f3ac71d27784caa390"} Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.044921 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77d99f96cd-fz4zl" event={"ID":"610e9c9c-cab0-4843-a120-5dde40363fd2","Type":"ContainerStarted","Data":"dc53aebae5c3c8a71d8c794757943e3ffcd34e1bfa28b3b68284b6d737e75e09"} Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.054878 4868 generic.go:334] "Generic (PLEG): container finished" podID="09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" containerID="b8a80047e73bb1f55aaaf76795af5f147f5d7b15b8532f53e7b071523d9e34a4" exitCode=137 Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.054911 4868 generic.go:334] "Generic (PLEG): container finished" podID="09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" containerID="b0b8ed13ae99e9296a2b0eadaacb695eae3ea49baffab2274a0b804636e28a2d" exitCode=137 Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.054991 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84b8d76fb9-9nh6j" event={"ID":"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20","Type":"ContainerDied","Data":"b8a80047e73bb1f55aaaf76795af5f147f5d7b15b8532f53e7b071523d9e34a4"} Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.055031 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84b8d76fb9-9nh6j" event={"ID":"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20","Type":"ContainerDied","Data":"b0b8ed13ae99e9296a2b0eadaacb695eae3ea49baffab2274a0b804636e28a2d"} Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.086681 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" event={"ID":"0a4cab6a-d367-4eff-aab1-5b20b99f855e","Type":"ContainerStarted","Data":"9bfb0d504fd052c0c647c5898498f58b39454aa2a610809af6075e47640ec893"} Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.100212 4868 generic.go:334] "Generic (PLEG): container finished" podID="8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" containerID="8aa42fb7693952e1198faa277e828f500f7f5c9df73038f7aed72d83081d0196" exitCode=137 Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.100251 4868 generic.go:334] "Generic (PLEG): container finished" podID="8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" containerID="3306b22dcff7cb03441971868a0d150d655932b2690fac92ca7e3d77390596f2" exitCode=137 Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.101016 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68d44bd6f7-s9nkw" event={"ID":"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a","Type":"ContainerDied","Data":"8aa42fb7693952e1198faa277e828f500f7f5c9df73038f7aed72d83081d0196"} Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.101083 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68d44bd6f7-s9nkw" event={"ID":"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a","Type":"ContainerDied","Data":"3306b22dcff7cb03441971868a0d150d655932b2690fac92ca7e3d77390596f2"} Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.199866 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5da0876e-2895-4dd0-baf2-27ca9e5a1c30" path="/var/lib/kubelet/pods/5da0876e-2895-4dd0-baf2-27ca9e5a1c30/volumes" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.519064 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.538654 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.604266 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-horizon-secret-key\") pod \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.604865 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-config-data\") pod \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.604937 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-scripts\") pod \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.605458 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-logs\") pod \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.605489 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zlbn\" (UniqueName: \"kubernetes.io/projected/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-kube-api-access-4zlbn\") pod \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\" (UID: \"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20\") " Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.612001 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-logs" (OuterVolumeSpecName: "logs") pod "09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" (UID: "09b8c0e9-1904-4fb9-8dc0-ab63eef59e20"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.612247 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.624701 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-kube-api-access-4zlbn" (OuterVolumeSpecName: "kube-api-access-4zlbn") pod "09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" (UID: "09b8c0e9-1904-4fb9-8dc0-ab63eef59e20"). InnerVolumeSpecName "kube-api-access-4zlbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.630436 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" (UID: "09b8c0e9-1904-4fb9-8dc0-ab63eef59e20"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.697715 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.700930 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-config-data" (OuterVolumeSpecName: "config-data") pod "09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" (UID: "09b8c0e9-1904-4fb9-8dc0-ab63eef59e20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.714743 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.714783 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zlbn\" (UniqueName: \"kubernetes.io/projected/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-kube-api-access-4zlbn\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.714794 4868 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.734351 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.765168 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-scripts" (OuterVolumeSpecName: "scripts") pod "09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" (UID: "09b8c0e9-1904-4fb9-8dc0-ab63eef59e20"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.817344 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l57k\" (UniqueName: \"kubernetes.io/projected/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-kube-api-access-4l57k\") pod \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.817414 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l98nx\" (UniqueName: \"kubernetes.io/projected/6cb38dd4-7c72-4b57-902f-67d10ccb033e-kube-api-access-l98nx\") pod \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.817476 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6cb38dd4-7c72-4b57-902f-67d10ccb033e-scripts\") pod \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.817494 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6cb38dd4-7c72-4b57-902f-67d10ccb033e-horizon-secret-key\") pod \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.817549 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-config-data\") pod \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.817564 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-scripts\") pod \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.817585 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-logs\") pod \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.817654 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-horizon-secret-key\") pod \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\" (UID: \"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a\") " Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.817686 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cb38dd4-7c72-4b57-902f-67d10ccb033e-logs\") pod \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.817759 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6cb38dd4-7c72-4b57-902f-67d10ccb033e-config-data\") pod \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\" (UID: \"6cb38dd4-7c72-4b57-902f-67d10ccb033e\") " Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.818680 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.823586 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-logs" (OuterVolumeSpecName: "logs") pod "8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" (UID: "8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.826765 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cb38dd4-7c72-4b57-902f-67d10ccb033e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6cb38dd4-7c72-4b57-902f-67d10ccb033e" (UID: "6cb38dd4-7c72-4b57-902f-67d10ccb033e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.830319 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cb38dd4-7c72-4b57-902f-67d10ccb033e-logs" (OuterVolumeSpecName: "logs") pod "6cb38dd4-7c72-4b57-902f-67d10ccb033e" (UID: "6cb38dd4-7c72-4b57-902f-67d10ccb033e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.841674 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" (UID: "8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.842577 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-kube-api-access-4l57k" (OuterVolumeSpecName: "kube-api-access-4l57k") pod "8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" (UID: "8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a"). InnerVolumeSpecName "kube-api-access-4l57k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.852424 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cb38dd4-7c72-4b57-902f-67d10ccb033e-kube-api-access-l98nx" (OuterVolumeSpecName: "kube-api-access-l98nx") pod "6cb38dd4-7c72-4b57-902f-67d10ccb033e" (UID: "6cb38dd4-7c72-4b57-902f-67d10ccb033e"). InnerVolumeSpecName "kube-api-access-l98nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.855942 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cb38dd4-7c72-4b57-902f-67d10ccb033e-scripts" (OuterVolumeSpecName: "scripts") pod "6cb38dd4-7c72-4b57-902f-67d10ccb033e" (UID: "6cb38dd4-7c72-4b57-902f-67d10ccb033e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.904172 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-config-data" (OuterVolumeSpecName: "config-data") pod "8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" (UID: "8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.913004 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-scripts" (OuterVolumeSpecName: "scripts") pod "8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" (UID: "8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.921300 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l57k\" (UniqueName: \"kubernetes.io/projected/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-kube-api-access-4l57k\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.921335 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l98nx\" (UniqueName: \"kubernetes.io/projected/6cb38dd4-7c72-4b57-902f-67d10ccb033e-kube-api-access-l98nx\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.921345 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6cb38dd4-7c72-4b57-902f-67d10ccb033e-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.921354 4868 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6cb38dd4-7c72-4b57-902f-67d10ccb033e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.921363 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.921372 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.921383 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.921391 4868 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.921399 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cb38dd4-7c72-4b57-902f-67d10ccb033e-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:06 crc kubenswrapper[4868]: I1201 17:44:06.930654 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cb38dd4-7c72-4b57-902f-67d10ccb033e-config-data" (OuterVolumeSpecName: "config-data") pod "6cb38dd4-7c72-4b57-902f-67d10ccb033e" (UID: "6cb38dd4-7c72-4b57-902f-67d10ccb033e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.023660 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6cb38dd4-7c72-4b57-902f-67d10ccb033e-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.139023 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c7c6bb95-lm4km" event={"ID":"6cb38dd4-7c72-4b57-902f-67d10ccb033e","Type":"ContainerDied","Data":"0be8aabd81eaafc0e733abc84716510eba4a535c7945b2cb31e366111ccfaccc"} Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.139099 4868 scope.go:117] "RemoveContainer" containerID="88b6a6d8f215885bd8a47e36bdf1a0a076588c47a0f63a65c967e873a4be3b43" Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.139268 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c7c6bb95-lm4km" Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.146527 4868 generic.go:334] "Generic (PLEG): container finished" podID="69c7a38b-c21b-46cd-a120-1d4cd1867941" containerID="6de604e3d047646f1da44a4148d0834d2d8d4dbaf250356cb700961f490323ec" exitCode=0 Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.146606 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" event={"ID":"69c7a38b-c21b-46cd-a120-1d4cd1867941","Type":"ContainerDied","Data":"6de604e3d047646f1da44a4148d0834d2d8d4dbaf250356cb700961f490323ec"} Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.181150 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-655b68fbf5-kxnvg" event={"ID":"6fdfb743-319d-4a56-8182-442da79d3ed2","Type":"ContainerStarted","Data":"4526ca5925d984ec0553b987577426b3b047efa1d3f133f191db2146bd1754fd"} Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.237466 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77d99f96cd-fz4zl" event={"ID":"610e9c9c-cab0-4843-a120-5dde40363fd2","Type":"ContainerStarted","Data":"af5f0b65aadb172220709d2f2448ab60880a7766a6665e0544c345185eac3354"} Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.238181 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.238243 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.258358 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-655b68fbf5-kxnvg" podStartSLOduration=6.51220813 podStartE2EDuration="17.258319406s" podCreationTimestamp="2025-12-01 17:43:50 +0000 UTC" firstStartedPulling="2025-12-01 17:43:52.648105151 +0000 UTC m=+1105.019215562" lastFinishedPulling="2025-12-01 17:44:03.394216427 +0000 UTC m=+1115.765326838" observedRunningTime="2025-12-01 17:44:07.209395236 +0000 UTC m=+1119.580505647" watchObservedRunningTime="2025-12-01 17:44:07.258319406 +0000 UTC m=+1119.629429817" Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.273375 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84b8d76fb9-9nh6j" Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.273995 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68c7c6bb95-lm4km"] Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.274081 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84b8d76fb9-9nh6j" event={"ID":"09b8c0e9-1904-4fb9-8dc0-ab63eef59e20","Type":"ContainerDied","Data":"b2e78c68534df4a916a3dee9d4b166e7c67a61e654f367911a6a6030e7db18b6"} Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.306886 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-68c7c6bb95-lm4km"] Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.312840 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-77d99f96cd-fz4zl" podStartSLOduration=13.312810629 podStartE2EDuration="13.312810629s" podCreationTimestamp="2025-12-01 17:43:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:44:07.289566742 +0000 UTC m=+1119.660677153" watchObservedRunningTime="2025-12-01 17:44:07.312810629 +0000 UTC m=+1119.683921040" Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.325173 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"57856c20-00b1-4106-aa7e-a935af8c2814","Type":"ContainerStarted","Data":"a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998"} Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.343352 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-84b8d76fb9-9nh6j"] Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.351693 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-84b8d76fb9-9nh6j"] Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.358903 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" event={"ID":"0a4cab6a-d367-4eff-aab1-5b20b99f855e","Type":"ContainerStarted","Data":"a647647dbf9613a90d9a8beca0ceac8b5616af2fa237e506fad4c1c1f9b4b600"} Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.390299 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68d44bd6f7-s9nkw" event={"ID":"8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a","Type":"ContainerDied","Data":"fb604386519d190da339dd0a1f9c434c12a9b9a2222d95e98f94c669e3ceafc1"} Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.390427 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68d44bd6f7-s9nkw" Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.395683 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-79b8fddb46-xqzzs" podStartSLOduration=9.071366715 podStartE2EDuration="17.39566144s" podCreationTimestamp="2025-12-01 17:43:50 +0000 UTC" firstStartedPulling="2025-12-01 17:43:52.521441779 +0000 UTC m=+1104.892552190" lastFinishedPulling="2025-12-01 17:44:00.845736494 +0000 UTC m=+1113.216846915" observedRunningTime="2025-12-01 17:44:07.386728196 +0000 UTC m=+1119.757838607" watchObservedRunningTime="2025-12-01 17:44:07.39566144 +0000 UTC m=+1119.766771851" Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.423170 4868 scope.go:117] "RemoveContainer" containerID="8ce65a2788466a9208f158286aa29b5856c5a7cf703453a0707de74b659b725b" Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.434591 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68d44bd6f7-s9nkw"] Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.454221 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-68d44bd6f7-s9nkw"] Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.478770 4868 scope.go:117] "RemoveContainer" containerID="b8a80047e73bb1f55aaaf76795af5f147f5d7b15b8532f53e7b071523d9e34a4" Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.744075 4868 scope.go:117] "RemoveContainer" containerID="b0b8ed13ae99e9296a2b0eadaacb695eae3ea49baffab2274a0b804636e28a2d" Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.792338 4868 scope.go:117] "RemoveContainer" containerID="8aa42fb7693952e1198faa277e828f500f7f5c9df73038f7aed72d83081d0196" Dec 01 17:44:07 crc kubenswrapper[4868]: I1201 17:44:07.983486 4868 scope.go:117] "RemoveContainer" containerID="3306b22dcff7cb03441971868a0d150d655932b2690fac92ca7e3d77390596f2" Dec 01 17:44:08 crc kubenswrapper[4868]: I1201 17:44:08.192624 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" path="/var/lib/kubelet/pods/09b8c0e9-1904-4fb9-8dc0-ab63eef59e20/volumes" Dec 01 17:44:08 crc kubenswrapper[4868]: I1201 17:44:08.193322 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cb38dd4-7c72-4b57-902f-67d10ccb033e" path="/var/lib/kubelet/pods/6cb38dd4-7c72-4b57-902f-67d10ccb033e/volumes" Dec 01 17:44:08 crc kubenswrapper[4868]: I1201 17:44:08.195486 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" path="/var/lib/kubelet/pods/8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a/volumes" Dec 01 17:44:08 crc kubenswrapper[4868]: I1201 17:44:08.414700 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" event={"ID":"69c7a38b-c21b-46cd-a120-1d4cd1867941","Type":"ContainerStarted","Data":"331aeb57154733165d0f6615693a045df1bd6043a1081c6b748d58c0c2bbd31a"} Dec 01 17:44:08 crc kubenswrapper[4868]: I1201 17:44:08.415831 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:08 crc kubenswrapper[4868]: I1201 17:44:08.428404 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"57856c20-00b1-4106-aa7e-a935af8c2814","Type":"ContainerStarted","Data":"3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5"} Dec 01 17:44:08 crc kubenswrapper[4868]: I1201 17:44:08.428646 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="57856c20-00b1-4106-aa7e-a935af8c2814" containerName="cinder-api-log" containerID="cri-o://a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998" gracePeriod=30 Dec 01 17:44:08 crc kubenswrapper[4868]: I1201 17:44:08.428677 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 17:44:08 crc kubenswrapper[4868]: I1201 17:44:08.428789 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="57856c20-00b1-4106-aa7e-a935af8c2814" containerName="cinder-api" containerID="cri-o://3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5" gracePeriod=30 Dec 01 17:44:08 crc kubenswrapper[4868]: I1201 17:44:08.441198 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" podStartSLOduration=6.441175993 podStartE2EDuration="6.441175993s" podCreationTimestamp="2025-12-01 17:44:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:44:08.437501903 +0000 UTC m=+1120.808612304" watchObservedRunningTime="2025-12-01 17:44:08.441175993 +0000 UTC m=+1120.812286404" Dec 01 17:44:08 crc kubenswrapper[4868]: I1201 17:44:08.486118 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.486083884 podStartE2EDuration="6.486083884s" podCreationTimestamp="2025-12-01 17:44:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:44:08.471509225 +0000 UTC m=+1120.842619636" watchObservedRunningTime="2025-12-01 17:44:08.486083884 +0000 UTC m=+1120.857194295" Dec 01 17:44:08 crc kubenswrapper[4868]: I1201 17:44:08.492811 4868 generic.go:334] "Generic (PLEG): container finished" podID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerID="e4775176a990584d72b3ad230e778ce5d004e058505e31837d7f84fd14e439f0" exitCode=0 Dec 01 17:44:08 crc kubenswrapper[4868]: I1201 17:44:08.492904 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4416e29-710d-4683-9c40-92ca09b0e0bd","Type":"ContainerDied","Data":"e4775176a990584d72b3ad230e778ce5d004e058505e31837d7f84fd14e439f0"} Dec 01 17:44:08 crc kubenswrapper[4868]: I1201 17:44:08.498208 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b311f726-7ee0-452d-be0e-545b36e38666","Type":"ContainerStarted","Data":"987d7a83b972e417801d6805790e5a717566def8f42cc766bc7be28a8f2c3738"} Dec 01 17:44:08 crc kubenswrapper[4868]: I1201 17:44:08.586455 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.320201 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.508394 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b311f726-7ee0-452d-be0e-545b36e38666","Type":"ContainerStarted","Data":"344c8e90dba9239c90abfe1825769c4923a82ae6466746500908cc9e3966cdad"} Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.510380 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-config-data\") pod \"57856c20-00b1-4106-aa7e-a935af8c2814\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.510503 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbdl2\" (UniqueName: \"kubernetes.io/projected/57856c20-00b1-4106-aa7e-a935af8c2814-kube-api-access-rbdl2\") pod \"57856c20-00b1-4106-aa7e-a935af8c2814\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.510634 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-scripts\") pod \"57856c20-00b1-4106-aa7e-a935af8c2814\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.510706 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-combined-ca-bundle\") pod \"57856c20-00b1-4106-aa7e-a935af8c2814\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.510734 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57856c20-00b1-4106-aa7e-a935af8c2814-logs\") pod \"57856c20-00b1-4106-aa7e-a935af8c2814\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.510787 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-config-data-custom\") pod \"57856c20-00b1-4106-aa7e-a935af8c2814\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.510935 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/57856c20-00b1-4106-aa7e-a935af8c2814-etc-machine-id\") pod \"57856c20-00b1-4106-aa7e-a935af8c2814\" (UID: \"57856c20-00b1-4106-aa7e-a935af8c2814\") " Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.511014 4868 generic.go:334] "Generic (PLEG): container finished" podID="57856c20-00b1-4106-aa7e-a935af8c2814" containerID="3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5" exitCode=0 Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.511037 4868 generic.go:334] "Generic (PLEG): container finished" podID="57856c20-00b1-4106-aa7e-a935af8c2814" containerID="a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998" exitCode=143 Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.511105 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.511133 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"57856c20-00b1-4106-aa7e-a935af8c2814","Type":"ContainerDied","Data":"3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5"} Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.511170 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"57856c20-00b1-4106-aa7e-a935af8c2814","Type":"ContainerDied","Data":"a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998"} Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.511183 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"57856c20-00b1-4106-aa7e-a935af8c2814","Type":"ContainerDied","Data":"d198b76742ac2e6df80489cd31fb920d60a48b75800ea7af18820a1247594585"} Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.511199 4868 scope.go:117] "RemoveContainer" containerID="3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.511471 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/57856c20-00b1-4106-aa7e-a935af8c2814-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "57856c20-00b1-4106-aa7e-a935af8c2814" (UID: "57856c20-00b1-4106-aa7e-a935af8c2814"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.511815 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57856c20-00b1-4106-aa7e-a935af8c2814-logs" (OuterVolumeSpecName: "logs") pod "57856c20-00b1-4106-aa7e-a935af8c2814" (UID: "57856c20-00b1-4106-aa7e-a935af8c2814"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.518274 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-scripts" (OuterVolumeSpecName: "scripts") pod "57856c20-00b1-4106-aa7e-a935af8c2814" (UID: "57856c20-00b1-4106-aa7e-a935af8c2814"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.523154 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57856c20-00b1-4106-aa7e-a935af8c2814-kube-api-access-rbdl2" (OuterVolumeSpecName: "kube-api-access-rbdl2") pod "57856c20-00b1-4106-aa7e-a935af8c2814" (UID: "57856c20-00b1-4106-aa7e-a935af8c2814"). InnerVolumeSpecName "kube-api-access-rbdl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.527444 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "57856c20-00b1-4106-aa7e-a935af8c2814" (UID: "57856c20-00b1-4106-aa7e-a935af8c2814"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.539404 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.899256481 podStartE2EDuration="7.539376561s" podCreationTimestamp="2025-12-01 17:44:02 +0000 UTC" firstStartedPulling="2025-12-01 17:44:04.707153249 +0000 UTC m=+1117.078263660" lastFinishedPulling="2025-12-01 17:44:06.347273329 +0000 UTC m=+1118.718383740" observedRunningTime="2025-12-01 17:44:09.526802156 +0000 UTC m=+1121.897912587" watchObservedRunningTime="2025-12-01 17:44:09.539376561 +0000 UTC m=+1121.910486972" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.544769 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57856c20-00b1-4106-aa7e-a935af8c2814" (UID: "57856c20-00b1-4106-aa7e-a935af8c2814"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.573172 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-config-data" (OuterVolumeSpecName: "config-data") pod "57856c20-00b1-4106-aa7e-a935af8c2814" (UID: "57856c20-00b1-4106-aa7e-a935af8c2814"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.613608 4868 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/57856c20-00b1-4106-aa7e-a935af8c2814-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.613650 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.613663 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbdl2\" (UniqueName: \"kubernetes.io/projected/57856c20-00b1-4106-aa7e-a935af8c2814-kube-api-access-rbdl2\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.613676 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.613691 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.613704 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57856c20-00b1-4106-aa7e-a935af8c2814-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.613715 4868 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57856c20-00b1-4106-aa7e-a935af8c2814-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.671522 4868 scope.go:117] "RemoveContainer" containerID="a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.694493 4868 scope.go:117] "RemoveContainer" containerID="3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5" Dec 01 17:44:09 crc kubenswrapper[4868]: E1201 17:44:09.695278 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5\": container with ID starting with 3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5 not found: ID does not exist" containerID="3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.695326 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5"} err="failed to get container status \"3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5\": rpc error: code = NotFound desc = could not find container \"3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5\": container with ID starting with 3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5 not found: ID does not exist" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.695357 4868 scope.go:117] "RemoveContainer" containerID="a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998" Dec 01 17:44:09 crc kubenswrapper[4868]: E1201 17:44:09.695677 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998\": container with ID starting with a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998 not found: ID does not exist" containerID="a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.695710 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998"} err="failed to get container status \"a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998\": rpc error: code = NotFound desc = could not find container \"a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998\": container with ID starting with a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998 not found: ID does not exist" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.695728 4868 scope.go:117] "RemoveContainer" containerID="3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.696199 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5"} err="failed to get container status \"3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5\": rpc error: code = NotFound desc = could not find container \"3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5\": container with ID starting with 3a22e179171e89fdef11a029a9a03c188a5c0b122e375b37455ba8db0a430fa5 not found: ID does not exist" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.696226 4868 scope.go:117] "RemoveContainer" containerID="a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.696460 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998"} err="failed to get container status \"a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998\": rpc error: code = NotFound desc = could not find container \"a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998\": container with ID starting with a448a7957cb80abf7d3561aa49a2adc7b4e654fe16a337b20f7a23d6ed26f998 not found: ID does not exist" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.785821 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5c9f554f77-pjbhr" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.874127 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-58c4c9c4cb-jdzxz"] Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.874487 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-58c4c9c4cb-jdzxz" podUID="c0cf4af0-a637-4d73-9f28-2ee6ea544584" containerName="neutron-api" containerID="cri-o://acbef70eb8c395a23a94cccf92bb42860a4718e29265b97fa27fbbb999c88e83" gracePeriod=30 Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.874670 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-58c4c9c4cb-jdzxz" podUID="c0cf4af0-a637-4d73-9f28-2ee6ea544584" containerName="neutron-httpd" containerID="cri-o://0886552474fe3abd4be9a720938c363213cb35a952612d81325fe20c40d49f01" gracePeriod=30 Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.889331 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.913734 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.923518 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 17:44:09 crc kubenswrapper[4868]: E1201 17:44:09.924004 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5da0876e-2895-4dd0-baf2-27ca9e5a1c30" containerName="init" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924024 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5da0876e-2895-4dd0-baf2-27ca9e5a1c30" containerName="init" Dec 01 17:44:09 crc kubenswrapper[4868]: E1201 17:44:09.924038 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" containerName="horizon" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924044 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" containerName="horizon" Dec 01 17:44:09 crc kubenswrapper[4868]: E1201 17:44:09.924053 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5da0876e-2895-4dd0-baf2-27ca9e5a1c30" containerName="dnsmasq-dns" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924060 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5da0876e-2895-4dd0-baf2-27ca9e5a1c30" containerName="dnsmasq-dns" Dec 01 17:44:09 crc kubenswrapper[4868]: E1201 17:44:09.924073 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57856c20-00b1-4106-aa7e-a935af8c2814" containerName="cinder-api-log" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924080 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="57856c20-00b1-4106-aa7e-a935af8c2814" containerName="cinder-api-log" Dec 01 17:44:09 crc kubenswrapper[4868]: E1201 17:44:09.924086 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57856c20-00b1-4106-aa7e-a935af8c2814" containerName="cinder-api" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924091 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="57856c20-00b1-4106-aa7e-a935af8c2814" containerName="cinder-api" Dec 01 17:44:09 crc kubenswrapper[4868]: E1201 17:44:09.924103 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cb38dd4-7c72-4b57-902f-67d10ccb033e" containerName="horizon" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924109 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cb38dd4-7c72-4b57-902f-67d10ccb033e" containerName="horizon" Dec 01 17:44:09 crc kubenswrapper[4868]: E1201 17:44:09.924121 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cb38dd4-7c72-4b57-902f-67d10ccb033e" containerName="horizon-log" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924127 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cb38dd4-7c72-4b57-902f-67d10ccb033e" containerName="horizon-log" Dec 01 17:44:09 crc kubenswrapper[4868]: E1201 17:44:09.924143 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" containerName="horizon-log" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924150 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" containerName="horizon-log" Dec 01 17:44:09 crc kubenswrapper[4868]: E1201 17:44:09.924164 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" containerName="horizon" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924171 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" containerName="horizon" Dec 01 17:44:09 crc kubenswrapper[4868]: E1201 17:44:09.924182 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" containerName="horizon-log" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924187 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" containerName="horizon-log" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924368 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5da0876e-2895-4dd0-baf2-27ca9e5a1c30" containerName="dnsmasq-dns" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924383 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" containerName="horizon-log" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924392 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" containerName="horizon-log" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924402 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="57856c20-00b1-4106-aa7e-a935af8c2814" containerName="cinder-api" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924411 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="57856c20-00b1-4106-aa7e-a935af8c2814" containerName="cinder-api-log" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924438 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cb38dd4-7c72-4b57-902f-67d10ccb033e" containerName="horizon-log" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924448 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cb38dd4-7c72-4b57-902f-67d10ccb033e" containerName="horizon" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924459 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b1e415d-9a19-4b21-8bd2-e1e2ee6ef11a" containerName="horizon" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.924466 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="09b8c0e9-1904-4fb9-8dc0-ab63eef59e20" containerName="horizon" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.925466 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.929437 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.931034 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.931391 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 01 17:44:09 crc kubenswrapper[4868]: I1201 17:44:09.949923 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.123417 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.123474 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.123590 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76a7855d-b030-49b6-a58f-4cdfc45d489a-logs\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.123650 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-config-data\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.123692 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flgfs\" (UniqueName: \"kubernetes.io/projected/76a7855d-b030-49b6-a58f-4cdfc45d489a-kube-api-access-flgfs\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.123731 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-public-tls-certs\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.123758 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-config-data-custom\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.123786 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-scripts\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.123810 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76a7855d-b030-49b6-a58f-4cdfc45d489a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.184236 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57856c20-00b1-4106-aa7e-a935af8c2814" path="/var/lib/kubelet/pods/57856c20-00b1-4106-aa7e-a935af8c2814/volumes" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.226492 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76a7855d-b030-49b6-a58f-4cdfc45d489a-logs\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.226670 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-config-data\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.226765 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flgfs\" (UniqueName: \"kubernetes.io/projected/76a7855d-b030-49b6-a58f-4cdfc45d489a-kube-api-access-flgfs\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.226923 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-public-tls-certs\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.227141 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-config-data-custom\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.227189 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-scripts\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.227225 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76a7855d-b030-49b6-a58f-4cdfc45d489a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.227233 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76a7855d-b030-49b6-a58f-4cdfc45d489a-logs\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.227283 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.227319 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.227668 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76a7855d-b030-49b6-a58f-4cdfc45d489a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.233214 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-config-data\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.233560 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-config-data-custom\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.237570 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-public-tls-certs\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.237901 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.238063 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.249833 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a7855d-b030-49b6-a58f-4cdfc45d489a-scripts\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.251346 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flgfs\" (UniqueName: \"kubernetes.io/projected/76a7855d-b030-49b6-a58f-4cdfc45d489a-kube-api-access-flgfs\") pod \"cinder-api-0\" (UID: \"76a7855d-b030-49b6-a58f-4cdfc45d489a\") " pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.254835 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.897736 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-77567b6ccb-79skk" Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.908849 4868 generic.go:334] "Generic (PLEG): container finished" podID="c0cf4af0-a637-4d73-9f28-2ee6ea544584" containerID="0886552474fe3abd4be9a720938c363213cb35a952612d81325fe20c40d49f01" exitCode=0 Dec 01 17:44:10 crc kubenswrapper[4868]: I1201 17:44:10.910493 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58c4c9c4cb-jdzxz" event={"ID":"c0cf4af0-a637-4d73-9f28-2ee6ea544584","Type":"ContainerDied","Data":"0886552474fe3abd4be9a720938c363213cb35a952612d81325fe20c40d49f01"} Dec 01 17:44:11 crc kubenswrapper[4868]: I1201 17:44:11.039212 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-97868dbf6-djbqg"] Dec 01 17:44:11 crc kubenswrapper[4868]: I1201 17:44:11.040245 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-97868dbf6-djbqg" podUID="fdb719ec-3812-4252-bdb9-1bc93aa35462" containerName="horizon-log" containerID="cri-o://98299ac37c9d398981045965e1ed7484fda45b7142252554239df2eab9ac0ba1" gracePeriod=30 Dec 01 17:44:11 crc kubenswrapper[4868]: I1201 17:44:11.042077 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-97868dbf6-djbqg" podUID="fdb719ec-3812-4252-bdb9-1bc93aa35462" containerName="horizon" containerID="cri-o://5dbf91a124b9ea9b19e7613ba78db4552ecf9b204c056b4966d3fd691eb80b32" gracePeriod=30 Dec 01 17:44:11 crc kubenswrapper[4868]: I1201 17:44:11.313282 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 17:44:11 crc kubenswrapper[4868]: W1201 17:44:11.326799 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76a7855d_b030_49b6_a58f_4cdfc45d489a.slice/crio-22e4834e6a760b8abf4378b9431a71e919d16cc7c4250fb2a4cc00f1f25a1082 WatchSource:0}: Error finding container 22e4834e6a760b8abf4378b9431a71e919d16cc7c4250fb2a4cc00f1f25a1082: Status 404 returned error can't find the container with id 22e4834e6a760b8abf4378b9431a71e919d16cc7c4250fb2a4cc00f1f25a1082 Dec 01 17:44:11 crc kubenswrapper[4868]: I1201 17:44:11.923402 4868 generic.go:334] "Generic (PLEG): container finished" podID="fdb719ec-3812-4252-bdb9-1bc93aa35462" containerID="5dbf91a124b9ea9b19e7613ba78db4552ecf9b204c056b4966d3fd691eb80b32" exitCode=0 Dec 01 17:44:11 crc kubenswrapper[4868]: I1201 17:44:11.923486 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-97868dbf6-djbqg" event={"ID":"fdb719ec-3812-4252-bdb9-1bc93aa35462","Type":"ContainerDied","Data":"5dbf91a124b9ea9b19e7613ba78db4552ecf9b204c056b4966d3fd691eb80b32"} Dec 01 17:44:11 crc kubenswrapper[4868]: I1201 17:44:11.925486 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"76a7855d-b030-49b6-a58f-4cdfc45d489a","Type":"ContainerStarted","Data":"22e4834e6a760b8abf4378b9431a71e919d16cc7c4250fb2a4cc00f1f25a1082"} Dec 01 17:44:11 crc kubenswrapper[4868]: I1201 17:44:11.975562 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:44:12 crc kubenswrapper[4868]: I1201 17:44:12.495111 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:44:12 crc kubenswrapper[4868]: I1201 17:44:12.650806 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7d8466bd7b-sr5pl" Dec 01 17:44:12 crc kubenswrapper[4868]: I1201 17:44:12.802408 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 17:44:12 crc kubenswrapper[4868]: I1201 17:44:12.823187 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:44:12 crc kubenswrapper[4868]: I1201 17:44:12.927253 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-nmgvr"] Dec 01 17:44:12 crc kubenswrapper[4868]: I1201 17:44:12.927494 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" podUID="b97acd99-8ef0-4c6d-a7d1-05f6c57450ad" containerName="dnsmasq-dns" containerID="cri-o://fc643f8472870f6b7005fd6b073a21091bb9d65a0757a0e3b19dbb4ab4dee38d" gracePeriod=10 Dec 01 17:44:12 crc kubenswrapper[4868]: I1201 17:44:12.968166 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"76a7855d-b030-49b6-a58f-4cdfc45d489a","Type":"ContainerStarted","Data":"f107a395ec4190095822761492255388f006ea47cf6d99d0a116b2174a56c6c3"} Dec 01 17:44:12 crc kubenswrapper[4868]: I1201 17:44:12.968227 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"76a7855d-b030-49b6-a58f-4cdfc45d489a","Type":"ContainerStarted","Data":"42542aa8aae631652934d5fcb2353ac3121bb0d934376058421247e03b858797"} Dec 01 17:44:12 crc kubenswrapper[4868]: I1201 17:44:12.969879 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.005977 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.005954776 podStartE2EDuration="4.005954776s" podCreationTimestamp="2025-12-01 17:44:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:44:13.000974379 +0000 UTC m=+1125.372084800" watchObservedRunningTime="2025-12-01 17:44:13.005954776 +0000 UTC m=+1125.377065187" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.181460 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.295554 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.652259 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.754290 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v46b9\" (UniqueName: \"kubernetes.io/projected/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-kube-api-access-v46b9\") pod \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.754390 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-dns-swift-storage-0\") pod \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.754558 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-dns-svc\") pod \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.754626 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-ovsdbserver-sb\") pod \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.754653 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-config\") pod \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.754742 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-ovsdbserver-nb\") pod \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\" (UID: \"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad\") " Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.762342 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-kube-api-access-v46b9" (OuterVolumeSpecName: "kube-api-access-v46b9") pod "b97acd99-8ef0-4c6d-a7d1-05f6c57450ad" (UID: "b97acd99-8ef0-4c6d-a7d1-05f6c57450ad"). InnerVolumeSpecName "kube-api-access-v46b9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.811110 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b97acd99-8ef0-4c6d-a7d1-05f6c57450ad" (UID: "b97acd99-8ef0-4c6d-a7d1-05f6c57450ad"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.812912 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b97acd99-8ef0-4c6d-a7d1-05f6c57450ad" (UID: "b97acd99-8ef0-4c6d-a7d1-05f6c57450ad"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.820316 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b97acd99-8ef0-4c6d-a7d1-05f6c57450ad" (UID: "b97acd99-8ef0-4c6d-a7d1-05f6c57450ad"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.827231 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b97acd99-8ef0-4c6d-a7d1-05f6c57450ad" (UID: "b97acd99-8ef0-4c6d-a7d1-05f6c57450ad"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.849748 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-config" (OuterVolumeSpecName: "config") pod "b97acd99-8ef0-4c6d-a7d1-05f6c57450ad" (UID: "b97acd99-8ef0-4c6d-a7d1-05f6c57450ad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.859105 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.859136 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v46b9\" (UniqueName: \"kubernetes.io/projected/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-kube-api-access-v46b9\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.859147 4868 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.859158 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.859168 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.859176 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.980570 4868 generic.go:334] "Generic (PLEG): container finished" podID="b97acd99-8ef0-4c6d-a7d1-05f6c57450ad" containerID="fc643f8472870f6b7005fd6b073a21091bb9d65a0757a0e3b19dbb4ab4dee38d" exitCode=0 Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.980794 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b311f726-7ee0-452d-be0e-545b36e38666" containerName="cinder-scheduler" containerID="cri-o://987d7a83b972e417801d6805790e5a717566def8f42cc766bc7be28a8f2c3738" gracePeriod=30 Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.981233 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.981478 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" event={"ID":"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad","Type":"ContainerDied","Data":"fc643f8472870f6b7005fd6b073a21091bb9d65a0757a0e3b19dbb4ab4dee38d"} Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.981512 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-nmgvr" event={"ID":"b97acd99-8ef0-4c6d-a7d1-05f6c57450ad","Type":"ContainerDied","Data":"a5e60ed02c6f31d22c070cd7dd23a00fda084129e355d86bad9267d90ea43340"} Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.981530 4868 scope.go:117] "RemoveContainer" containerID="fc643f8472870f6b7005fd6b073a21091bb9d65a0757a0e3b19dbb4ab4dee38d" Dec 01 17:44:13 crc kubenswrapper[4868]: I1201 17:44:13.983058 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b311f726-7ee0-452d-be0e-545b36e38666" containerName="probe" containerID="cri-o://344c8e90dba9239c90abfe1825769c4923a82ae6466746500908cc9e3966cdad" gracePeriod=30 Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.009581 4868 scope.go:117] "RemoveContainer" containerID="704f8ef1fdfbea100f187f24fe378da5ea11adf7cf689d1e1b17cc8f6c1fe6cc" Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.027139 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-nmgvr"] Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.035688 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-nmgvr"] Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.037489 4868 scope.go:117] "RemoveContainer" containerID="fc643f8472870f6b7005fd6b073a21091bb9d65a0757a0e3b19dbb4ab4dee38d" Dec 01 17:44:14 crc kubenswrapper[4868]: E1201 17:44:14.038021 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc643f8472870f6b7005fd6b073a21091bb9d65a0757a0e3b19dbb4ab4dee38d\": container with ID starting with fc643f8472870f6b7005fd6b073a21091bb9d65a0757a0e3b19dbb4ab4dee38d not found: ID does not exist" containerID="fc643f8472870f6b7005fd6b073a21091bb9d65a0757a0e3b19dbb4ab4dee38d" Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.038062 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc643f8472870f6b7005fd6b073a21091bb9d65a0757a0e3b19dbb4ab4dee38d"} err="failed to get container status \"fc643f8472870f6b7005fd6b073a21091bb9d65a0757a0e3b19dbb4ab4dee38d\": rpc error: code = NotFound desc = could not find container \"fc643f8472870f6b7005fd6b073a21091bb9d65a0757a0e3b19dbb4ab4dee38d\": container with ID starting with fc643f8472870f6b7005fd6b073a21091bb9d65a0757a0e3b19dbb4ab4dee38d not found: ID does not exist" Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.038092 4868 scope.go:117] "RemoveContainer" containerID="704f8ef1fdfbea100f187f24fe378da5ea11adf7cf689d1e1b17cc8f6c1fe6cc" Dec 01 17:44:14 crc kubenswrapper[4868]: E1201 17:44:14.038620 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"704f8ef1fdfbea100f187f24fe378da5ea11adf7cf689d1e1b17cc8f6c1fe6cc\": container with ID starting with 704f8ef1fdfbea100f187f24fe378da5ea11adf7cf689d1e1b17cc8f6c1fe6cc not found: ID does not exist" containerID="704f8ef1fdfbea100f187f24fe378da5ea11adf7cf689d1e1b17cc8f6c1fe6cc" Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.038644 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"704f8ef1fdfbea100f187f24fe378da5ea11adf7cf689d1e1b17cc8f6c1fe6cc"} err="failed to get container status \"704f8ef1fdfbea100f187f24fe378da5ea11adf7cf689d1e1b17cc8f6c1fe6cc\": rpc error: code = NotFound desc = could not find container \"704f8ef1fdfbea100f187f24fe378da5ea11adf7cf689d1e1b17cc8f6c1fe6cc\": container with ID starting with 704f8ef1fdfbea100f187f24fe378da5ea11adf7cf689d1e1b17cc8f6c1fe6cc not found: ID does not exist" Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.166088 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-77d99f96cd-fz4zl" Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.182167 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b97acd99-8ef0-4c6d-a7d1-05f6c57450ad" path="/var/lib/kubelet/pods/b97acd99-8ef0-4c6d-a7d1-05f6c57450ad/volumes" Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.251541 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6969f75cfd-5nchg"] Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.251798 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6969f75cfd-5nchg" podUID="c87b6efa-9ca1-4d35-9ff5-51e466b925d2" containerName="barbican-api-log" containerID="cri-o://ab3d83823b8b622d9d6fb5190ead32db337340e8c8e2e8cf3a7f154666001f13" gracePeriod=30 Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.251935 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6969f75cfd-5nchg" podUID="c87b6efa-9ca1-4d35-9ff5-51e466b925d2" containerName="barbican-api" containerID="cri-o://b733d416cd8a7287bce1bfd5b1b79fa10fe77f54f603d5c47f9d4ec32e984853" gracePeriod=30 Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.991624 4868 generic.go:334] "Generic (PLEG): container finished" podID="c87b6efa-9ca1-4d35-9ff5-51e466b925d2" containerID="ab3d83823b8b622d9d6fb5190ead32db337340e8c8e2e8cf3a7f154666001f13" exitCode=143 Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.991725 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6969f75cfd-5nchg" event={"ID":"c87b6efa-9ca1-4d35-9ff5-51e466b925d2","Type":"ContainerDied","Data":"ab3d83823b8b622d9d6fb5190ead32db337340e8c8e2e8cf3a7f154666001f13"} Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.995043 4868 generic.go:334] "Generic (PLEG): container finished" podID="b311f726-7ee0-452d-be0e-545b36e38666" containerID="344c8e90dba9239c90abfe1825769c4923a82ae6466746500908cc9e3966cdad" exitCode=0 Dec 01 17:44:14 crc kubenswrapper[4868]: I1201 17:44:14.995082 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b311f726-7ee0-452d-be0e-545b36e38666","Type":"ContainerDied","Data":"344c8e90dba9239c90abfe1825769c4923a82ae6466746500908cc9e3966cdad"} Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.009559 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.009629 4868 generic.go:334] "Generic (PLEG): container finished" podID="b311f726-7ee0-452d-be0e-545b36e38666" containerID="987d7a83b972e417801d6805790e5a717566def8f42cc766bc7be28a8f2c3738" exitCode=0 Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.009659 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b311f726-7ee0-452d-be0e-545b36e38666","Type":"ContainerDied","Data":"987d7a83b972e417801d6805790e5a717566def8f42cc766bc7be28a8f2c3738"} Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.010126 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b311f726-7ee0-452d-be0e-545b36e38666","Type":"ContainerDied","Data":"dcb11f7c381ef006b26266a9347e2fba80bbf60089e9abc217f6afe2b0233e09"} Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.010157 4868 scope.go:117] "RemoveContainer" containerID="344c8e90dba9239c90abfe1825769c4923a82ae6466746500908cc9e3966cdad" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.040096 4868 scope.go:117] "RemoveContainer" containerID="987d7a83b972e417801d6805790e5a717566def8f42cc766bc7be28a8f2c3738" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.080902 4868 scope.go:117] "RemoveContainer" containerID="344c8e90dba9239c90abfe1825769c4923a82ae6466746500908cc9e3966cdad" Dec 01 17:44:16 crc kubenswrapper[4868]: E1201 17:44:16.081650 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"344c8e90dba9239c90abfe1825769c4923a82ae6466746500908cc9e3966cdad\": container with ID starting with 344c8e90dba9239c90abfe1825769c4923a82ae6466746500908cc9e3966cdad not found: ID does not exist" containerID="344c8e90dba9239c90abfe1825769c4923a82ae6466746500908cc9e3966cdad" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.081713 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"344c8e90dba9239c90abfe1825769c4923a82ae6466746500908cc9e3966cdad"} err="failed to get container status \"344c8e90dba9239c90abfe1825769c4923a82ae6466746500908cc9e3966cdad\": rpc error: code = NotFound desc = could not find container \"344c8e90dba9239c90abfe1825769c4923a82ae6466746500908cc9e3966cdad\": container with ID starting with 344c8e90dba9239c90abfe1825769c4923a82ae6466746500908cc9e3966cdad not found: ID does not exist" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.081749 4868 scope.go:117] "RemoveContainer" containerID="987d7a83b972e417801d6805790e5a717566def8f42cc766bc7be28a8f2c3738" Dec 01 17:44:16 crc kubenswrapper[4868]: E1201 17:44:16.082394 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"987d7a83b972e417801d6805790e5a717566def8f42cc766bc7be28a8f2c3738\": container with ID starting with 987d7a83b972e417801d6805790e5a717566def8f42cc766bc7be28a8f2c3738 not found: ID does not exist" containerID="987d7a83b972e417801d6805790e5a717566def8f42cc766bc7be28a8f2c3738" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.082429 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"987d7a83b972e417801d6805790e5a717566def8f42cc766bc7be28a8f2c3738"} err="failed to get container status \"987d7a83b972e417801d6805790e5a717566def8f42cc766bc7be28a8f2c3738\": rpc error: code = NotFound desc = could not find container \"987d7a83b972e417801d6805790e5a717566def8f42cc766bc7be28a8f2c3738\": container with ID starting with 987d7a83b972e417801d6805790e5a717566def8f42cc766bc7be28a8f2c3738 not found: ID does not exist" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.104677 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-config-data\") pod \"b311f726-7ee0-452d-be0e-545b36e38666\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.105156 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-combined-ca-bundle\") pod \"b311f726-7ee0-452d-be0e-545b36e38666\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.105234 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b311f726-7ee0-452d-be0e-545b36e38666-etc-machine-id\") pod \"b311f726-7ee0-452d-be0e-545b36e38666\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.105309 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-scripts\") pod \"b311f726-7ee0-452d-be0e-545b36e38666\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.105358 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pmqd\" (UniqueName: \"kubernetes.io/projected/b311f726-7ee0-452d-be0e-545b36e38666-kube-api-access-5pmqd\") pod \"b311f726-7ee0-452d-be0e-545b36e38666\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.105362 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b311f726-7ee0-452d-be0e-545b36e38666-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b311f726-7ee0-452d-be0e-545b36e38666" (UID: "b311f726-7ee0-452d-be0e-545b36e38666"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.106098 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-config-data-custom\") pod \"b311f726-7ee0-452d-be0e-545b36e38666\" (UID: \"b311f726-7ee0-452d-be0e-545b36e38666\") " Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.106551 4868 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b311f726-7ee0-452d-be0e-545b36e38666-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.114096 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b311f726-7ee0-452d-be0e-545b36e38666-kube-api-access-5pmqd" (OuterVolumeSpecName: "kube-api-access-5pmqd") pod "b311f726-7ee0-452d-be0e-545b36e38666" (UID: "b311f726-7ee0-452d-be0e-545b36e38666"). InnerVolumeSpecName "kube-api-access-5pmqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.114231 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b311f726-7ee0-452d-be0e-545b36e38666" (UID: "b311f726-7ee0-452d-be0e-545b36e38666"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.115035 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-scripts" (OuterVolumeSpecName: "scripts") pod "b311f726-7ee0-452d-be0e-545b36e38666" (UID: "b311f726-7ee0-452d-be0e-545b36e38666"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.165150 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b311f726-7ee0-452d-be0e-545b36e38666" (UID: "b311f726-7ee0-452d-be0e-545b36e38666"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.209214 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.209249 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.209259 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pmqd\" (UniqueName: \"kubernetes.io/projected/b311f726-7ee0-452d-be0e-545b36e38666-kube-api-access-5pmqd\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.209271 4868 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.257086 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-config-data" (OuterVolumeSpecName: "config-data") pod "b311f726-7ee0-452d-be0e-545b36e38666" (UID: "b311f726-7ee0-452d-be0e-545b36e38666"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.311563 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b311f726-7ee0-452d-be0e-545b36e38666-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.530826 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.619578 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn6tz\" (UniqueName: \"kubernetes.io/projected/c0cf4af0-a637-4d73-9f28-2ee6ea544584-kube-api-access-fn6tz\") pod \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.619843 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-ovndb-tls-certs\") pod \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.619925 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-config\") pod \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.619993 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-combined-ca-bundle\") pod \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.620028 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-httpd-config\") pod \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\" (UID: \"c0cf4af0-a637-4d73-9f28-2ee6ea544584\") " Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.626580 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "c0cf4af0-a637-4d73-9f28-2ee6ea544584" (UID: "c0cf4af0-a637-4d73-9f28-2ee6ea544584"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.628266 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0cf4af0-a637-4d73-9f28-2ee6ea544584-kube-api-access-fn6tz" (OuterVolumeSpecName: "kube-api-access-fn6tz") pod "c0cf4af0-a637-4d73-9f28-2ee6ea544584" (UID: "c0cf4af0-a637-4d73-9f28-2ee6ea544584"). InnerVolumeSpecName "kube-api-access-fn6tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.685490 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0cf4af0-a637-4d73-9f28-2ee6ea544584" (UID: "c0cf4af0-a637-4d73-9f28-2ee6ea544584"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.691088 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-config" (OuterVolumeSpecName: "config") pod "c0cf4af0-a637-4d73-9f28-2ee6ea544584" (UID: "c0cf4af0-a637-4d73-9f28-2ee6ea544584"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.715045 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "c0cf4af0-a637-4d73-9f28-2ee6ea544584" (UID: "c0cf4af0-a637-4d73-9f28-2ee6ea544584"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.722456 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.722489 4868 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.722501 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn6tz\" (UniqueName: \"kubernetes.io/projected/c0cf4af0-a637-4d73-9f28-2ee6ea544584-kube-api-access-fn6tz\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.722513 4868 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:16 crc kubenswrapper[4868]: I1201 17:44:16.722522 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c0cf4af0-a637-4d73-9f28-2ee6ea544584-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.029583 4868 generic.go:334] "Generic (PLEG): container finished" podID="c0cf4af0-a637-4d73-9f28-2ee6ea544584" containerID="acbef70eb8c395a23a94cccf92bb42860a4718e29265b97fa27fbbb999c88e83" exitCode=0 Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.029744 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58c4c9c4cb-jdzxz" event={"ID":"c0cf4af0-a637-4d73-9f28-2ee6ea544584","Type":"ContainerDied","Data":"acbef70eb8c395a23a94cccf92bb42860a4718e29265b97fa27fbbb999c88e83"} Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.029757 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58c4c9c4cb-jdzxz" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.029809 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58c4c9c4cb-jdzxz" event={"ID":"c0cf4af0-a637-4d73-9f28-2ee6ea544584","Type":"ContainerDied","Data":"f9d94714ddd30af11be90141bd15ece7a3531d6c95627f1e4eeff3a1fb143d49"} Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.029837 4868 scope.go:117] "RemoveContainer" containerID="0886552474fe3abd4be9a720938c363213cb35a952612d81325fe20c40d49f01" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.035195 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.063900 4868 scope.go:117] "RemoveContainer" containerID="acbef70eb8c395a23a94cccf92bb42860a4718e29265b97fa27fbbb999c88e83" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.075171 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-58c4c9c4cb-jdzxz"] Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.086798 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-58c4c9c4cb-jdzxz"] Dec 01 17:44:17 crc kubenswrapper[4868]: E1201 17:44:17.101588 4868 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb311f726_7ee0_452d_be0e_545b36e38666.slice\": RecentStats: unable to find data in memory cache]" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.117473 4868 scope.go:117] "RemoveContainer" containerID="0886552474fe3abd4be9a720938c363213cb35a952612d81325fe20c40d49f01" Dec 01 17:44:17 crc kubenswrapper[4868]: E1201 17:44:17.118334 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0886552474fe3abd4be9a720938c363213cb35a952612d81325fe20c40d49f01\": container with ID starting with 0886552474fe3abd4be9a720938c363213cb35a952612d81325fe20c40d49f01 not found: ID does not exist" containerID="0886552474fe3abd4be9a720938c363213cb35a952612d81325fe20c40d49f01" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.118371 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0886552474fe3abd4be9a720938c363213cb35a952612d81325fe20c40d49f01"} err="failed to get container status \"0886552474fe3abd4be9a720938c363213cb35a952612d81325fe20c40d49f01\": rpc error: code = NotFound desc = could not find container \"0886552474fe3abd4be9a720938c363213cb35a952612d81325fe20c40d49f01\": container with ID starting with 0886552474fe3abd4be9a720938c363213cb35a952612d81325fe20c40d49f01 not found: ID does not exist" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.118418 4868 scope.go:117] "RemoveContainer" containerID="acbef70eb8c395a23a94cccf92bb42860a4718e29265b97fa27fbbb999c88e83" Dec 01 17:44:17 crc kubenswrapper[4868]: E1201 17:44:17.118726 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acbef70eb8c395a23a94cccf92bb42860a4718e29265b97fa27fbbb999c88e83\": container with ID starting with acbef70eb8c395a23a94cccf92bb42860a4718e29265b97fa27fbbb999c88e83 not found: ID does not exist" containerID="acbef70eb8c395a23a94cccf92bb42860a4718e29265b97fa27fbbb999c88e83" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.118791 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acbef70eb8c395a23a94cccf92bb42860a4718e29265b97fa27fbbb999c88e83"} err="failed to get container status \"acbef70eb8c395a23a94cccf92bb42860a4718e29265b97fa27fbbb999c88e83\": rpc error: code = NotFound desc = could not find container \"acbef70eb8c395a23a94cccf92bb42860a4718e29265b97fa27fbbb999c88e83\": container with ID starting with acbef70eb8c395a23a94cccf92bb42860a4718e29265b97fa27fbbb999c88e83 not found: ID does not exist" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.119274 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.130457 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.139940 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 17:44:17 crc kubenswrapper[4868]: E1201 17:44:17.140345 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b97acd99-8ef0-4c6d-a7d1-05f6c57450ad" containerName="init" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.140364 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b97acd99-8ef0-4c6d-a7d1-05f6c57450ad" containerName="init" Dec 01 17:44:17 crc kubenswrapper[4868]: E1201 17:44:17.140377 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0cf4af0-a637-4d73-9f28-2ee6ea544584" containerName="neutron-httpd" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.140386 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0cf4af0-a637-4d73-9f28-2ee6ea544584" containerName="neutron-httpd" Dec 01 17:44:17 crc kubenswrapper[4868]: E1201 17:44:17.140415 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b311f726-7ee0-452d-be0e-545b36e38666" containerName="cinder-scheduler" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.140421 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b311f726-7ee0-452d-be0e-545b36e38666" containerName="cinder-scheduler" Dec 01 17:44:17 crc kubenswrapper[4868]: E1201 17:44:17.140432 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b97acd99-8ef0-4c6d-a7d1-05f6c57450ad" containerName="dnsmasq-dns" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.140438 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b97acd99-8ef0-4c6d-a7d1-05f6c57450ad" containerName="dnsmasq-dns" Dec 01 17:44:17 crc kubenswrapper[4868]: E1201 17:44:17.140449 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b311f726-7ee0-452d-be0e-545b36e38666" containerName="probe" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.140455 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b311f726-7ee0-452d-be0e-545b36e38666" containerName="probe" Dec 01 17:44:17 crc kubenswrapper[4868]: E1201 17:44:17.140474 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0cf4af0-a637-4d73-9f28-2ee6ea544584" containerName="neutron-api" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.140480 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0cf4af0-a637-4d73-9f28-2ee6ea544584" containerName="neutron-api" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.140634 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0cf4af0-a637-4d73-9f28-2ee6ea544584" containerName="neutron-httpd" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.140645 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0cf4af0-a637-4d73-9f28-2ee6ea544584" containerName="neutron-api" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.140654 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="b311f726-7ee0-452d-be0e-545b36e38666" containerName="cinder-scheduler" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.140672 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="b97acd99-8ef0-4c6d-a7d1-05f6c57450ad" containerName="dnsmasq-dns" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.140687 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="b311f726-7ee0-452d-be0e-545b36e38666" containerName="probe" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.141671 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.148260 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.151627 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.231373 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8a69c3d-3924-4160-87ec-17a00572a260-config-data\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.231444 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8a69c3d-3924-4160-87ec-17a00572a260-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.231523 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6b5f\" (UniqueName: \"kubernetes.io/projected/d8a69c3d-3924-4160-87ec-17a00572a260-kube-api-access-z6b5f\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.231605 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8a69c3d-3924-4160-87ec-17a00572a260-scripts\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.231628 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a69c3d-3924-4160-87ec-17a00572a260-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.231770 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d8a69c3d-3924-4160-87ec-17a00572a260-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.332755 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d8a69c3d-3924-4160-87ec-17a00572a260-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.332873 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d8a69c3d-3924-4160-87ec-17a00572a260-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.333866 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8a69c3d-3924-4160-87ec-17a00572a260-config-data\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.333920 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8a69c3d-3924-4160-87ec-17a00572a260-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.333994 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6b5f\" (UniqueName: \"kubernetes.io/projected/d8a69c3d-3924-4160-87ec-17a00572a260-kube-api-access-z6b5f\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.334084 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8a69c3d-3924-4160-87ec-17a00572a260-scripts\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.334119 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a69c3d-3924-4160-87ec-17a00572a260-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.338778 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8a69c3d-3924-4160-87ec-17a00572a260-config-data\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.339301 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8a69c3d-3924-4160-87ec-17a00572a260-scripts\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.341293 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8a69c3d-3924-4160-87ec-17a00572a260-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.354071 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a69c3d-3924-4160-87ec-17a00572a260-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.356231 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6b5f\" (UniqueName: \"kubernetes.io/projected/d8a69c3d-3924-4160-87ec-17a00572a260-kube-api-access-z6b5f\") pod \"cinder-scheduler-0\" (UID: \"d8a69c3d-3924-4160-87ec-17a00572a260\") " pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.462573 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.690713 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6969f75cfd-5nchg" podUID="c87b6efa-9ca1-4d35-9ff5-51e466b925d2" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:53242->10.217.0.160:9311: read: connection reset by peer" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.690735 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6969f75cfd-5nchg" podUID="c87b6efa-9ca1-4d35-9ff5-51e466b925d2" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:53226->10.217.0.160:9311: read: connection reset by peer" Dec 01 17:44:17 crc kubenswrapper[4868]: I1201 17:44:17.930031 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.046705 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d8a69c3d-3924-4160-87ec-17a00572a260","Type":"ContainerStarted","Data":"fa73ced916fb7587c6b574832905d2a0723e9d15841faed8ee82d56c42ca1df5"} Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.050398 4868 generic.go:334] "Generic (PLEG): container finished" podID="c87b6efa-9ca1-4d35-9ff5-51e466b925d2" containerID="b733d416cd8a7287bce1bfd5b1b79fa10fe77f54f603d5c47f9d4ec32e984853" exitCode=0 Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.050459 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6969f75cfd-5nchg" event={"ID":"c87b6efa-9ca1-4d35-9ff5-51e466b925d2","Type":"ContainerDied","Data":"b733d416cd8a7287bce1bfd5b1b79fa10fe77f54f603d5c47f9d4ec32e984853"} Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.148044 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.216883 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b311f726-7ee0-452d-be0e-545b36e38666" path="/var/lib/kubelet/pods/b311f726-7ee0-452d-be0e-545b36e38666/volumes" Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.218167 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0cf4af0-a637-4d73-9f28-2ee6ea544584" path="/var/lib/kubelet/pods/c0cf4af0-a637-4d73-9f28-2ee6ea544584/volumes" Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.251030 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-combined-ca-bundle\") pod \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.251140 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-config-data-custom\") pod \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.251182 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-config-data\") pod \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.251270 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-logs\") pod \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.251394 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx5wn\" (UniqueName: \"kubernetes.io/projected/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-kube-api-access-xx5wn\") pod \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\" (UID: \"c87b6efa-9ca1-4d35-9ff5-51e466b925d2\") " Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.253569 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-logs" (OuterVolumeSpecName: "logs") pod "c87b6efa-9ca1-4d35-9ff5-51e466b925d2" (UID: "c87b6efa-9ca1-4d35-9ff5-51e466b925d2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.257872 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c87b6efa-9ca1-4d35-9ff5-51e466b925d2" (UID: "c87b6efa-9ca1-4d35-9ff5-51e466b925d2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.258660 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-kube-api-access-xx5wn" (OuterVolumeSpecName: "kube-api-access-xx5wn") pod "c87b6efa-9ca1-4d35-9ff5-51e466b925d2" (UID: "c87b6efa-9ca1-4d35-9ff5-51e466b925d2"). InnerVolumeSpecName "kube-api-access-xx5wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.285270 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c87b6efa-9ca1-4d35-9ff5-51e466b925d2" (UID: "c87b6efa-9ca1-4d35-9ff5-51e466b925d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.307288 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-config-data" (OuterVolumeSpecName: "config-data") pod "c87b6efa-9ca1-4d35-9ff5-51e466b925d2" (UID: "c87b6efa-9ca1-4d35-9ff5-51e466b925d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.358444 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.358531 4868 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.358547 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.358557 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:18 crc kubenswrapper[4868]: I1201 17:44:18.358568 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx5wn\" (UniqueName: \"kubernetes.io/projected/c87b6efa-9ca1-4d35-9ff5-51e466b925d2-kube-api-access-xx5wn\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:19 crc kubenswrapper[4868]: I1201 17:44:19.065139 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d8a69c3d-3924-4160-87ec-17a00572a260","Type":"ContainerStarted","Data":"1f383ca7c919bafa4c45f9e5037a55693b7cd89bf0567b4365527fe1af26d122"} Dec 01 17:44:19 crc kubenswrapper[4868]: I1201 17:44:19.065597 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d8a69c3d-3924-4160-87ec-17a00572a260","Type":"ContainerStarted","Data":"24625cecf7fdfc97ab34170921e4c8ba0c255ab0108cf246203636dfd08429c0"} Dec 01 17:44:19 crc kubenswrapper[4868]: I1201 17:44:19.068633 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6969f75cfd-5nchg" event={"ID":"c87b6efa-9ca1-4d35-9ff5-51e466b925d2","Type":"ContainerDied","Data":"4b959695ec42dc9c5a1d7300b44624e3dd91cc3eae500f40687235a453d64ff7"} Dec 01 17:44:19 crc kubenswrapper[4868]: I1201 17:44:19.068671 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6969f75cfd-5nchg" Dec 01 17:44:19 crc kubenswrapper[4868]: I1201 17:44:19.068692 4868 scope.go:117] "RemoveContainer" containerID="b733d416cd8a7287bce1bfd5b1b79fa10fe77f54f603d5c47f9d4ec32e984853" Dec 01 17:44:19 crc kubenswrapper[4868]: I1201 17:44:19.090614 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.090595291 podStartE2EDuration="2.090595291s" podCreationTimestamp="2025-12-01 17:44:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:44:19.089684516 +0000 UTC m=+1131.460794927" watchObservedRunningTime="2025-12-01 17:44:19.090595291 +0000 UTC m=+1131.461705702" Dec 01 17:44:19 crc kubenswrapper[4868]: I1201 17:44:19.102627 4868 scope.go:117] "RemoveContainer" containerID="ab3d83823b8b622d9d6fb5190ead32db337340e8c8e2e8cf3a7f154666001f13" Dec 01 17:44:19 crc kubenswrapper[4868]: I1201 17:44:19.117704 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6969f75cfd-5nchg"] Dec 01 17:44:19 crc kubenswrapper[4868]: I1201 17:44:19.125201 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6969f75cfd-5nchg"] Dec 01 17:44:20 crc kubenswrapper[4868]: I1201 17:44:20.185278 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c87b6efa-9ca1-4d35-9ff5-51e466b925d2" path="/var/lib/kubelet/pods/c87b6efa-9ca1-4d35-9ff5-51e466b925d2/volumes" Dec 01 17:44:22 crc kubenswrapper[4868]: I1201 17:44:22.013032 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 01 17:44:22 crc kubenswrapper[4868]: I1201 17:44:22.463745 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 17:44:22 crc kubenswrapper[4868]: I1201 17:44:22.835232 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5476cd6644-rgwc7" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.824471 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 17:44:24 crc kubenswrapper[4868]: E1201 17:44:24.825173 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c87b6efa-9ca1-4d35-9ff5-51e466b925d2" containerName="barbican-api" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.825187 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="c87b6efa-9ca1-4d35-9ff5-51e466b925d2" containerName="barbican-api" Dec 01 17:44:24 crc kubenswrapper[4868]: E1201 17:44:24.825198 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c87b6efa-9ca1-4d35-9ff5-51e466b925d2" containerName="barbican-api-log" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.825205 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="c87b6efa-9ca1-4d35-9ff5-51e466b925d2" containerName="barbican-api-log" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.825433 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="c87b6efa-9ca1-4d35-9ff5-51e466b925d2" containerName="barbican-api" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.825452 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="c87b6efa-9ca1-4d35-9ff5-51e466b925d2" containerName="barbican-api-log" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.827750 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.831271 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.831856 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-dlqg5" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.842534 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.866703 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.890733 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hpr8\" (UniqueName: \"kubernetes.io/projected/c8e4121e-9a64-47f9-9335-3784e46968e9-kube-api-access-5hpr8\") pod \"openstackclient\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " pod="openstack/openstackclient" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.891150 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c8e4121e-9a64-47f9-9335-3784e46968e9-openstack-config-secret\") pod \"openstackclient\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " pod="openstack/openstackclient" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.891298 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8e4121e-9a64-47f9-9335-3784e46968e9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " pod="openstack/openstackclient" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.891649 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c8e4121e-9a64-47f9-9335-3784e46968e9-openstack-config\") pod \"openstackclient\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " pod="openstack/openstackclient" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.994024 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c8e4121e-9a64-47f9-9335-3784e46968e9-openstack-config-secret\") pod \"openstackclient\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " pod="openstack/openstackclient" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.994384 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8e4121e-9a64-47f9-9335-3784e46968e9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " pod="openstack/openstackclient" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.994650 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c8e4121e-9a64-47f9-9335-3784e46968e9-openstack-config\") pod \"openstackclient\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " pod="openstack/openstackclient" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.995013 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hpr8\" (UniqueName: \"kubernetes.io/projected/c8e4121e-9a64-47f9-9335-3784e46968e9-kube-api-access-5hpr8\") pod \"openstackclient\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " pod="openstack/openstackclient" Dec 01 17:44:24 crc kubenswrapper[4868]: I1201 17:44:24.995883 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c8e4121e-9a64-47f9-9335-3784e46968e9-openstack-config\") pod \"openstackclient\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.002071 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8e4121e-9a64-47f9-9335-3784e46968e9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.003441 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c8e4121e-9a64-47f9-9335-3784e46968e9-openstack-config-secret\") pod \"openstackclient\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.013634 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hpr8\" (UniqueName: \"kubernetes.io/projected/c8e4121e-9a64-47f9-9335-3784e46968e9-kube-api-access-5hpr8\") pod \"openstackclient\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.161349 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.240923 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.249017 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.314814 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.330859 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.331709 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 17:44:25 crc kubenswrapper[4868]: E1201 17:44:25.399651 4868 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 01 17:44:25 crc kubenswrapper[4868]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_c8e4121e-9a64-47f9-9335-3784e46968e9_0(3ebd9691c9663984006194dbf271e8b8206cad11ed136044b7da3bbc9158b09f): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"3ebd9691c9663984006194dbf271e8b8206cad11ed136044b7da3bbc9158b09f" Netns:"/var/run/netns/43bcba6d-ba54-4e2b-bd08-cbbd2c2623b0" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=3ebd9691c9663984006194dbf271e8b8206cad11ed136044b7da3bbc9158b09f;K8S_POD_UID=c8e4121e-9a64-47f9-9335-3784e46968e9" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/c8e4121e-9a64-47f9-9335-3784e46968e9]: expected pod UID "c8e4121e-9a64-47f9-9335-3784e46968e9" but got "3f3f5b1c-3ed5-4d79-b298-474f483b2c4a" from Kube API Dec 01 17:44:25 crc kubenswrapper[4868]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 01 17:44:25 crc kubenswrapper[4868]: > Dec 01 17:44:25 crc kubenswrapper[4868]: E1201 17:44:25.400070 4868 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 01 17:44:25 crc kubenswrapper[4868]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_c8e4121e-9a64-47f9-9335-3784e46968e9_0(3ebd9691c9663984006194dbf271e8b8206cad11ed136044b7da3bbc9158b09f): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"3ebd9691c9663984006194dbf271e8b8206cad11ed136044b7da3bbc9158b09f" Netns:"/var/run/netns/43bcba6d-ba54-4e2b-bd08-cbbd2c2623b0" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=3ebd9691c9663984006194dbf271e8b8206cad11ed136044b7da3bbc9158b09f;K8S_POD_UID=c8e4121e-9a64-47f9-9335-3784e46968e9" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/c8e4121e-9a64-47f9-9335-3784e46968e9]: expected pod UID "c8e4121e-9a64-47f9-9335-3784e46968e9" but got "3f3f5b1c-3ed5-4d79-b298-474f483b2c4a" from Kube API Dec 01 17:44:25 crc kubenswrapper[4868]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 01 17:44:25 crc kubenswrapper[4868]: > pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.406862 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsz9w\" (UniqueName: \"kubernetes.io/projected/3f3f5b1c-3ed5-4d79-b298-474f483b2c4a-kube-api-access-bsz9w\") pod \"openstackclient\" (UID: \"3f3f5b1c-3ed5-4d79-b298-474f483b2c4a\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.407375 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3f3f5b1c-3ed5-4d79-b298-474f483b2c4a-openstack-config\") pod \"openstackclient\" (UID: \"3f3f5b1c-3ed5-4d79-b298-474f483b2c4a\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.407508 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3f3f5b1c-3ed5-4d79-b298-474f483b2c4a-openstack-config-secret\") pod \"openstackclient\" (UID: \"3f3f5b1c-3ed5-4d79-b298-474f483b2c4a\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.407733 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f3f5b1c-3ed5-4d79-b298-474f483b2c4a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3f3f5b1c-3ed5-4d79-b298-474f483b2c4a\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.510304 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsz9w\" (UniqueName: \"kubernetes.io/projected/3f3f5b1c-3ed5-4d79-b298-474f483b2c4a-kube-api-access-bsz9w\") pod \"openstackclient\" (UID: \"3f3f5b1c-3ed5-4d79-b298-474f483b2c4a\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.510685 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3f3f5b1c-3ed5-4d79-b298-474f483b2c4a-openstack-config\") pod \"openstackclient\" (UID: \"3f3f5b1c-3ed5-4d79-b298-474f483b2c4a\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.510714 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3f3f5b1c-3ed5-4d79-b298-474f483b2c4a-openstack-config-secret\") pod \"openstackclient\" (UID: \"3f3f5b1c-3ed5-4d79-b298-474f483b2c4a\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.510841 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f3f5b1c-3ed5-4d79-b298-474f483b2c4a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3f3f5b1c-3ed5-4d79-b298-474f483b2c4a\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.511683 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3f3f5b1c-3ed5-4d79-b298-474f483b2c4a-openstack-config\") pod \"openstackclient\" (UID: \"3f3f5b1c-3ed5-4d79-b298-474f483b2c4a\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.517848 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3f3f5b1c-3ed5-4d79-b298-474f483b2c4a-openstack-config-secret\") pod \"openstackclient\" (UID: \"3f3f5b1c-3ed5-4d79-b298-474f483b2c4a\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.517921 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f3f5b1c-3ed5-4d79-b298-474f483b2c4a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3f3f5b1c-3ed5-4d79-b298-474f483b2c4a\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.527020 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsz9w\" (UniqueName: \"kubernetes.io/projected/3f3f5b1c-3ed5-4d79-b298-474f483b2c4a-kube-api-access-bsz9w\") pod \"openstackclient\" (UID: \"3f3f5b1c-3ed5-4d79-b298-474f483b2c4a\") " pod="openstack/openstackclient" Dec 01 17:44:25 crc kubenswrapper[4868]: I1201 17:44:25.657927 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.105295 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.137060 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.143989 4868 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="c8e4121e-9a64-47f9-9335-3784e46968e9" podUID="3f3f5b1c-3ed5-4d79-b298-474f483b2c4a" Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.179807 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.230707 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c8e4121e-9a64-47f9-9335-3784e46968e9-openstack-config-secret\") pod \"c8e4121e-9a64-47f9-9335-3784e46968e9\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.232758 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hpr8\" (UniqueName: \"kubernetes.io/projected/c8e4121e-9a64-47f9-9335-3784e46968e9-kube-api-access-5hpr8\") pod \"c8e4121e-9a64-47f9-9335-3784e46968e9\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.232819 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c8e4121e-9a64-47f9-9335-3784e46968e9-openstack-config\") pod \"c8e4121e-9a64-47f9-9335-3784e46968e9\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.232857 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8e4121e-9a64-47f9-9335-3784e46968e9-combined-ca-bundle\") pod \"c8e4121e-9a64-47f9-9335-3784e46968e9\" (UID: \"c8e4121e-9a64-47f9-9335-3784e46968e9\") " Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.233495 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8e4121e-9a64-47f9-9335-3784e46968e9-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "c8e4121e-9a64-47f9-9335-3784e46968e9" (UID: "c8e4121e-9a64-47f9-9335-3784e46968e9"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.233969 4868 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c8e4121e-9a64-47f9-9335-3784e46968e9-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.241779 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8e4121e-9a64-47f9-9335-3784e46968e9-kube-api-access-5hpr8" (OuterVolumeSpecName: "kube-api-access-5hpr8") pod "c8e4121e-9a64-47f9-9335-3784e46968e9" (UID: "c8e4121e-9a64-47f9-9335-3784e46968e9"). InnerVolumeSpecName "kube-api-access-5hpr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.242105 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8e4121e-9a64-47f9-9335-3784e46968e9-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "c8e4121e-9a64-47f9-9335-3784e46968e9" (UID: "c8e4121e-9a64-47f9-9335-3784e46968e9"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.247939 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8e4121e-9a64-47f9-9335-3784e46968e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8e4121e-9a64-47f9-9335-3784e46968e9" (UID: "c8e4121e-9a64-47f9-9335-3784e46968e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.335953 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hpr8\" (UniqueName: \"kubernetes.io/projected/c8e4121e-9a64-47f9-9335-3784e46968e9-kube-api-access-5hpr8\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.335992 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8e4121e-9a64-47f9-9335-3784e46968e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:26 crc kubenswrapper[4868]: I1201 17:44:26.336001 4868 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c8e4121e-9a64-47f9-9335-3784e46968e9-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:27 crc kubenswrapper[4868]: I1201 17:44:27.152795 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3f3f5b1c-3ed5-4d79-b298-474f483b2c4a","Type":"ContainerStarted","Data":"a1d977d671ac41388b65b357cdce7711b19f1717262048c77a0ea10f5709971a"} Dec 01 17:44:27 crc kubenswrapper[4868]: I1201 17:44:27.152817 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 17:44:27 crc kubenswrapper[4868]: I1201 17:44:27.166968 4868 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="c8e4121e-9a64-47f9-9335-3784e46968e9" podUID="3f3f5b1c-3ed5-4d79-b298-474f483b2c4a" Dec 01 17:44:27 crc kubenswrapper[4868]: E1201 17:44:27.365653 4868 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8e4121e_9a64_47f9_9335_3784e46968e9.slice\": RecentStats: unable to find data in memory cache]" Dec 01 17:44:27 crc kubenswrapper[4868]: I1201 17:44:27.746246 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.192774 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8e4121e-9a64-47f9-9335-3784e46968e9" path="/var/lib/kubelet/pods/c8e4121e-9a64-47f9-9335-3784e46968e9/volumes" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.851010 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-86984fb57-t5hzr"] Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.853607 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.866579 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.866622 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.866781 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.881337 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-86984fb57-t5hzr"] Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.884736 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a427adaa-2060-4639-bd6d-e53b8fb00357-log-httpd\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.884782 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a427adaa-2060-4639-bd6d-e53b8fb00357-config-data\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.884863 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a427adaa-2060-4639-bd6d-e53b8fb00357-internal-tls-certs\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.884884 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a427adaa-2060-4639-bd6d-e53b8fb00357-run-httpd\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.884910 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a427adaa-2060-4639-bd6d-e53b8fb00357-public-tls-certs\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.884964 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8b9w\" (UniqueName: \"kubernetes.io/projected/a427adaa-2060-4639-bd6d-e53b8fb00357-kube-api-access-n8b9w\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.884994 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a427adaa-2060-4639-bd6d-e53b8fb00357-combined-ca-bundle\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.885021 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a427adaa-2060-4639-bd6d-e53b8fb00357-etc-swift\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.986184 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a427adaa-2060-4639-bd6d-e53b8fb00357-config-data\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.986333 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a427adaa-2060-4639-bd6d-e53b8fb00357-internal-tls-certs\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.986360 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a427adaa-2060-4639-bd6d-e53b8fb00357-run-httpd\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.986395 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a427adaa-2060-4639-bd6d-e53b8fb00357-public-tls-certs\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.986450 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8b9w\" (UniqueName: \"kubernetes.io/projected/a427adaa-2060-4639-bd6d-e53b8fb00357-kube-api-access-n8b9w\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.986505 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a427adaa-2060-4639-bd6d-e53b8fb00357-combined-ca-bundle\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.986545 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a427adaa-2060-4639-bd6d-e53b8fb00357-etc-swift\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.986576 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a427adaa-2060-4639-bd6d-e53b8fb00357-log-httpd\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.987026 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a427adaa-2060-4639-bd6d-e53b8fb00357-log-httpd\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.987205 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a427adaa-2060-4639-bd6d-e53b8fb00357-run-httpd\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.992703 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a427adaa-2060-4639-bd6d-e53b8fb00357-public-tls-certs\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.993232 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a427adaa-2060-4639-bd6d-e53b8fb00357-etc-swift\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.995073 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a427adaa-2060-4639-bd6d-e53b8fb00357-combined-ca-bundle\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:28 crc kubenswrapper[4868]: I1201 17:44:28.996268 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a427adaa-2060-4639-bd6d-e53b8fb00357-config-data\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:29 crc kubenswrapper[4868]: I1201 17:44:29.004782 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8b9w\" (UniqueName: \"kubernetes.io/projected/a427adaa-2060-4639-bd6d-e53b8fb00357-kube-api-access-n8b9w\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:29 crc kubenswrapper[4868]: I1201 17:44:29.022673 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a427adaa-2060-4639-bd6d-e53b8fb00357-internal-tls-certs\") pod \"swift-proxy-86984fb57-t5hzr\" (UID: \"a427adaa-2060-4639-bd6d-e53b8fb00357\") " pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:29 crc kubenswrapper[4868]: I1201 17:44:29.198961 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:29 crc kubenswrapper[4868]: I1201 17:44:29.800879 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-86984fb57-t5hzr"] Dec 01 17:44:30 crc kubenswrapper[4868]: I1201 17:44:30.193808 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-86984fb57-t5hzr" event={"ID":"a427adaa-2060-4639-bd6d-e53b8fb00357","Type":"ContainerStarted","Data":"c899d344d0548cb9f9d4d3effcda54031c30842f93107c2b38c0ea9e2565516c"} Dec 01 17:44:30 crc kubenswrapper[4868]: I1201 17:44:30.194127 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-86984fb57-t5hzr" event={"ID":"a427adaa-2060-4639-bd6d-e53b8fb00357","Type":"ContainerStarted","Data":"d60b6e2c5167b2aa856a550a4ec0b8f50492466b48c4cb7d94f8115112429af5"} Dec 01 17:44:31 crc kubenswrapper[4868]: I1201 17:44:31.204762 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-86984fb57-t5hzr" event={"ID":"a427adaa-2060-4639-bd6d-e53b8fb00357","Type":"ContainerStarted","Data":"0c288d942c51a108f4deb18be03fa48a3cdc86221cccd75f312feec0a1f25363"} Dec 01 17:44:31 crc kubenswrapper[4868]: I1201 17:44:31.205194 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:31 crc kubenswrapper[4868]: I1201 17:44:31.238474 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-86984fb57-t5hzr" podStartSLOduration=3.238451434 podStartE2EDuration="3.238451434s" podCreationTimestamp="2025-12-01 17:44:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:44:31.223566677 +0000 UTC m=+1143.594677088" watchObservedRunningTime="2025-12-01 17:44:31.238451434 +0000 UTC m=+1143.609561845" Dec 01 17:44:31 crc kubenswrapper[4868]: I1201 17:44:31.812709 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 17:44:31 crc kubenswrapper[4868]: I1201 17:44:31.813550 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f6e2a84b-c570-44fa-9eba-1fbe11b761eb" containerName="glance-log" containerID="cri-o://b9832a951b9d366bd17c83326764a4e46db1979d6f50477ef7fd5e2ea671ac9d" gracePeriod=30 Dec 01 17:44:31 crc kubenswrapper[4868]: I1201 17:44:31.813657 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f6e2a84b-c570-44fa-9eba-1fbe11b761eb" containerName="glance-httpd" containerID="cri-o://277c8c1ea3ef347a4fac8ddbe7c0add43671c594e88b66a7eee55784ddd140eb" gracePeriod=30 Dec 01 17:44:32 crc kubenswrapper[4868]: I1201 17:44:32.217836 4868 generic.go:334] "Generic (PLEG): container finished" podID="f6e2a84b-c570-44fa-9eba-1fbe11b761eb" containerID="b9832a951b9d366bd17c83326764a4e46db1979d6f50477ef7fd5e2ea671ac9d" exitCode=143 Dec 01 17:44:32 crc kubenswrapper[4868]: I1201 17:44:32.217894 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f6e2a84b-c570-44fa-9eba-1fbe11b761eb","Type":"ContainerDied","Data":"b9832a951b9d366bd17c83326764a4e46db1979d6f50477ef7fd5e2ea671ac9d"} Dec 01 17:44:32 crc kubenswrapper[4868]: I1201 17:44:32.218098 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:32 crc kubenswrapper[4868]: I1201 17:44:32.550576 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 17:44:32 crc kubenswrapper[4868]: I1201 17:44:32.551094 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="168bf194-0fdb-4919-89d9-cf91e0395c21" containerName="glance-httpd" containerID="cri-o://cf1b13312b06ed5a6f43881a3f655dcda1ff81353748e85d8a3e963732db1509" gracePeriod=30 Dec 01 17:44:32 crc kubenswrapper[4868]: I1201 17:44:32.551283 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="168bf194-0fdb-4919-89d9-cf91e0395c21" containerName="glance-log" containerID="cri-o://5dd285127895f6cb2cb5b9ca97f148b3d95e701f72c41175af3d807c0a595a59" gracePeriod=30 Dec 01 17:44:33 crc kubenswrapper[4868]: I1201 17:44:33.230233 4868 generic.go:334] "Generic (PLEG): container finished" podID="168bf194-0fdb-4919-89d9-cf91e0395c21" containerID="5dd285127895f6cb2cb5b9ca97f148b3d95e701f72c41175af3d807c0a595a59" exitCode=143 Dec 01 17:44:33 crc kubenswrapper[4868]: I1201 17:44:33.230424 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"168bf194-0fdb-4919-89d9-cf91e0395c21","Type":"ContainerDied","Data":"5dd285127895f6cb2cb5b9ca97f148b3d95e701f72c41175af3d807c0a595a59"} Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.581908 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-xgtmd"] Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.588328 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xgtmd" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.591737 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xgtmd"] Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.679306 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-xfwp6"] Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.681600 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xfwp6" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.699760 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-xfwp6"] Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.714016 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-bc8a-account-create-update-ht62v"] Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.715621 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bc8a-account-create-update-ht62v" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.718302 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.732786 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kq7k\" (UniqueName: \"kubernetes.io/projected/b801bf8a-a39f-4537-aa83-06e958c31e02-kube-api-access-8kq7k\") pod \"nova-cell0-db-create-xfwp6\" (UID: \"b801bf8a-a39f-4537-aa83-06e958c31e02\") " pod="openstack/nova-cell0-db-create-xfwp6" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.732915 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5vjj\" (UniqueName: \"kubernetes.io/projected/31d8af7d-f71e-42be-ab59-237f2d505b1f-kube-api-access-k5vjj\") pod \"nova-api-db-create-xgtmd\" (UID: \"31d8af7d-f71e-42be-ab59-237f2d505b1f\") " pod="openstack/nova-api-db-create-xgtmd" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.733051 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b801bf8a-a39f-4537-aa83-06e958c31e02-operator-scripts\") pod \"nova-cell0-db-create-xfwp6\" (UID: \"b801bf8a-a39f-4537-aa83-06e958c31e02\") " pod="openstack/nova-cell0-db-create-xfwp6" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.733107 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31d8af7d-f71e-42be-ab59-237f2d505b1f-operator-scripts\") pod \"nova-api-db-create-xgtmd\" (UID: \"31d8af7d-f71e-42be-ab59-237f2d505b1f\") " pod="openstack/nova-api-db-create-xgtmd" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.739633 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-bc8a-account-create-update-ht62v"] Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.835587 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b801bf8a-a39f-4537-aa83-06e958c31e02-operator-scripts\") pod \"nova-cell0-db-create-xfwp6\" (UID: \"b801bf8a-a39f-4537-aa83-06e958c31e02\") " pod="openstack/nova-cell0-db-create-xfwp6" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.835649 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31d8af7d-f71e-42be-ab59-237f2d505b1f-operator-scripts\") pod \"nova-api-db-create-xgtmd\" (UID: \"31d8af7d-f71e-42be-ab59-237f2d505b1f\") " pod="openstack/nova-api-db-create-xgtmd" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.835681 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6m29\" (UniqueName: \"kubernetes.io/projected/a996536c-2c87-4c80-90d6-c6758c109af1-kube-api-access-d6m29\") pod \"nova-api-bc8a-account-create-update-ht62v\" (UID: \"a996536c-2c87-4c80-90d6-c6758c109af1\") " pod="openstack/nova-api-bc8a-account-create-update-ht62v" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.835759 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a996536c-2c87-4c80-90d6-c6758c109af1-operator-scripts\") pod \"nova-api-bc8a-account-create-update-ht62v\" (UID: \"a996536c-2c87-4c80-90d6-c6758c109af1\") " pod="openstack/nova-api-bc8a-account-create-update-ht62v" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.835780 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kq7k\" (UniqueName: \"kubernetes.io/projected/b801bf8a-a39f-4537-aa83-06e958c31e02-kube-api-access-8kq7k\") pod \"nova-cell0-db-create-xfwp6\" (UID: \"b801bf8a-a39f-4537-aa83-06e958c31e02\") " pod="openstack/nova-cell0-db-create-xfwp6" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.835802 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5vjj\" (UniqueName: \"kubernetes.io/projected/31d8af7d-f71e-42be-ab59-237f2d505b1f-kube-api-access-k5vjj\") pod \"nova-api-db-create-xgtmd\" (UID: \"31d8af7d-f71e-42be-ab59-237f2d505b1f\") " pod="openstack/nova-api-db-create-xgtmd" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.836793 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b801bf8a-a39f-4537-aa83-06e958c31e02-operator-scripts\") pod \"nova-cell0-db-create-xfwp6\" (UID: \"b801bf8a-a39f-4537-aa83-06e958c31e02\") " pod="openstack/nova-cell0-db-create-xfwp6" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.837058 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31d8af7d-f71e-42be-ab59-237f2d505b1f-operator-scripts\") pod \"nova-api-db-create-xgtmd\" (UID: \"31d8af7d-f71e-42be-ab59-237f2d505b1f\") " pod="openstack/nova-api-db-create-xgtmd" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.878105 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5vjj\" (UniqueName: \"kubernetes.io/projected/31d8af7d-f71e-42be-ab59-237f2d505b1f-kube-api-access-k5vjj\") pod \"nova-api-db-create-xgtmd\" (UID: \"31d8af7d-f71e-42be-ab59-237f2d505b1f\") " pod="openstack/nova-api-db-create-xgtmd" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.890811 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kq7k\" (UniqueName: \"kubernetes.io/projected/b801bf8a-a39f-4537-aa83-06e958c31e02-kube-api-access-8kq7k\") pod \"nova-cell0-db-create-xfwp6\" (UID: \"b801bf8a-a39f-4537-aa83-06e958c31e02\") " pod="openstack/nova-cell0-db-create-xfwp6" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.906004 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-3252-account-create-update-jrq2v"] Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.907337 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3252-account-create-update-jrq2v" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.916384 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.921308 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-wh5cc"] Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.923154 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wh5cc" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.925256 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xgtmd" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.937001 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-wh5cc"] Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.938119 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a996536c-2c87-4c80-90d6-c6758c109af1-operator-scripts\") pod \"nova-api-bc8a-account-create-update-ht62v\" (UID: \"a996536c-2c87-4c80-90d6-c6758c109af1\") " pod="openstack/nova-api-bc8a-account-create-update-ht62v" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.938235 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6m29\" (UniqueName: \"kubernetes.io/projected/a996536c-2c87-4c80-90d6-c6758c109af1-kube-api-access-d6m29\") pod \"nova-api-bc8a-account-create-update-ht62v\" (UID: \"a996536c-2c87-4c80-90d6-c6758c109af1\") " pod="openstack/nova-api-bc8a-account-create-update-ht62v" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.939405 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a996536c-2c87-4c80-90d6-c6758c109af1-operator-scripts\") pod \"nova-api-bc8a-account-create-update-ht62v\" (UID: \"a996536c-2c87-4c80-90d6-c6758c109af1\") " pod="openstack/nova-api-bc8a-account-create-update-ht62v" Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.948028 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3252-account-create-update-jrq2v"] Dec 01 17:44:34 crc kubenswrapper[4868]: I1201 17:44:34.969607 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6m29\" (UniqueName: \"kubernetes.io/projected/a996536c-2c87-4c80-90d6-c6758c109af1-kube-api-access-d6m29\") pod \"nova-api-bc8a-account-create-update-ht62v\" (UID: \"a996536c-2c87-4c80-90d6-c6758c109af1\") " pod="openstack/nova-api-bc8a-account-create-update-ht62v" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.000437 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xfwp6" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.005265 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.139:3000/\": dial tcp 10.217.0.139:3000: connect: connection refused" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.036894 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bc8a-account-create-update-ht62v" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.040548 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fknrm\" (UniqueName: \"kubernetes.io/projected/288e3235-6690-4c67-9e4f-2735be76d249-kube-api-access-fknrm\") pod \"nova-cell1-db-create-wh5cc\" (UID: \"288e3235-6690-4c67-9e4f-2735be76d249\") " pod="openstack/nova-cell1-db-create-wh5cc" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.040779 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzcdr\" (UniqueName: \"kubernetes.io/projected/e5951a13-5101-4956-bcf6-345fab68ea75-kube-api-access-vzcdr\") pod \"nova-cell0-3252-account-create-update-jrq2v\" (UID: \"e5951a13-5101-4956-bcf6-345fab68ea75\") " pod="openstack/nova-cell0-3252-account-create-update-jrq2v" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.040845 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5951a13-5101-4956-bcf6-345fab68ea75-operator-scripts\") pod \"nova-cell0-3252-account-create-update-jrq2v\" (UID: \"e5951a13-5101-4956-bcf6-345fab68ea75\") " pod="openstack/nova-cell0-3252-account-create-update-jrq2v" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.040872 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/288e3235-6690-4c67-9e4f-2735be76d249-operator-scripts\") pod \"nova-cell1-db-create-wh5cc\" (UID: \"288e3235-6690-4c67-9e4f-2735be76d249\") " pod="openstack/nova-cell1-db-create-wh5cc" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.082206 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-0c2b-account-create-update-gqzxq"] Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.083784 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0c2b-account-create-update-gqzxq" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.087634 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.094174 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-0c2b-account-create-update-gqzxq"] Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.142649 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvnfv\" (UniqueName: \"kubernetes.io/projected/441fc0fe-f3fe-4e94-abd8-0d126b289042-kube-api-access-dvnfv\") pod \"nova-cell1-0c2b-account-create-update-gqzxq\" (UID: \"441fc0fe-f3fe-4e94-abd8-0d126b289042\") " pod="openstack/nova-cell1-0c2b-account-create-update-gqzxq" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.142722 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fknrm\" (UniqueName: \"kubernetes.io/projected/288e3235-6690-4c67-9e4f-2735be76d249-kube-api-access-fknrm\") pod \"nova-cell1-db-create-wh5cc\" (UID: \"288e3235-6690-4c67-9e4f-2735be76d249\") " pod="openstack/nova-cell1-db-create-wh5cc" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.142804 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzcdr\" (UniqueName: \"kubernetes.io/projected/e5951a13-5101-4956-bcf6-345fab68ea75-kube-api-access-vzcdr\") pod \"nova-cell0-3252-account-create-update-jrq2v\" (UID: \"e5951a13-5101-4956-bcf6-345fab68ea75\") " pod="openstack/nova-cell0-3252-account-create-update-jrq2v" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.142866 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5951a13-5101-4956-bcf6-345fab68ea75-operator-scripts\") pod \"nova-cell0-3252-account-create-update-jrq2v\" (UID: \"e5951a13-5101-4956-bcf6-345fab68ea75\") " pod="openstack/nova-cell0-3252-account-create-update-jrq2v" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.142933 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/288e3235-6690-4c67-9e4f-2735be76d249-operator-scripts\") pod \"nova-cell1-db-create-wh5cc\" (UID: \"288e3235-6690-4c67-9e4f-2735be76d249\") " pod="openstack/nova-cell1-db-create-wh5cc" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.143019 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/441fc0fe-f3fe-4e94-abd8-0d126b289042-operator-scripts\") pod \"nova-cell1-0c2b-account-create-update-gqzxq\" (UID: \"441fc0fe-f3fe-4e94-abd8-0d126b289042\") " pod="openstack/nova-cell1-0c2b-account-create-update-gqzxq" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.144183 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5951a13-5101-4956-bcf6-345fab68ea75-operator-scripts\") pod \"nova-cell0-3252-account-create-update-jrq2v\" (UID: \"e5951a13-5101-4956-bcf6-345fab68ea75\") " pod="openstack/nova-cell0-3252-account-create-update-jrq2v" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.144466 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/288e3235-6690-4c67-9e4f-2735be76d249-operator-scripts\") pod \"nova-cell1-db-create-wh5cc\" (UID: \"288e3235-6690-4c67-9e4f-2735be76d249\") " pod="openstack/nova-cell1-db-create-wh5cc" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.169698 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fknrm\" (UniqueName: \"kubernetes.io/projected/288e3235-6690-4c67-9e4f-2735be76d249-kube-api-access-fknrm\") pod \"nova-cell1-db-create-wh5cc\" (UID: \"288e3235-6690-4c67-9e4f-2735be76d249\") " pod="openstack/nova-cell1-db-create-wh5cc" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.203354 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzcdr\" (UniqueName: \"kubernetes.io/projected/e5951a13-5101-4956-bcf6-345fab68ea75-kube-api-access-vzcdr\") pod \"nova-cell0-3252-account-create-update-jrq2v\" (UID: \"e5951a13-5101-4956-bcf6-345fab68ea75\") " pod="openstack/nova-cell0-3252-account-create-update-jrq2v" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.244878 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/441fc0fe-f3fe-4e94-abd8-0d126b289042-operator-scripts\") pod \"nova-cell1-0c2b-account-create-update-gqzxq\" (UID: \"441fc0fe-f3fe-4e94-abd8-0d126b289042\") " pod="openstack/nova-cell1-0c2b-account-create-update-gqzxq" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.245027 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvnfv\" (UniqueName: \"kubernetes.io/projected/441fc0fe-f3fe-4e94-abd8-0d126b289042-kube-api-access-dvnfv\") pod \"nova-cell1-0c2b-account-create-update-gqzxq\" (UID: \"441fc0fe-f3fe-4e94-abd8-0d126b289042\") " pod="openstack/nova-cell1-0c2b-account-create-update-gqzxq" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.247418 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/441fc0fe-f3fe-4e94-abd8-0d126b289042-operator-scripts\") pod \"nova-cell1-0c2b-account-create-update-gqzxq\" (UID: \"441fc0fe-f3fe-4e94-abd8-0d126b289042\") " pod="openstack/nova-cell1-0c2b-account-create-update-gqzxq" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.265451 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvnfv\" (UniqueName: \"kubernetes.io/projected/441fc0fe-f3fe-4e94-abd8-0d126b289042-kube-api-access-dvnfv\") pod \"nova-cell1-0c2b-account-create-update-gqzxq\" (UID: \"441fc0fe-f3fe-4e94-abd8-0d126b289042\") " pod="openstack/nova-cell1-0c2b-account-create-update-gqzxq" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.267349 4868 generic.go:334] "Generic (PLEG): container finished" podID="f6e2a84b-c570-44fa-9eba-1fbe11b761eb" containerID="277c8c1ea3ef347a4fac8ddbe7c0add43671c594e88b66a7eee55784ddd140eb" exitCode=0 Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.267412 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f6e2a84b-c570-44fa-9eba-1fbe11b761eb","Type":"ContainerDied","Data":"277c8c1ea3ef347a4fac8ddbe7c0add43671c594e88b66a7eee55784ddd140eb"} Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.270724 4868 generic.go:334] "Generic (PLEG): container finished" podID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerID="7024addc687087043e4fb44941d29311179976a32ceaac0d20843ca1623961eb" exitCode=137 Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.270779 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4416e29-710d-4683-9c40-92ca09b0e0bd","Type":"ContainerDied","Data":"7024addc687087043e4fb44941d29311179976a32ceaac0d20843ca1623961eb"} Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.456272 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3252-account-create-update-jrq2v" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.467164 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wh5cc" Dec 01 17:44:35 crc kubenswrapper[4868]: I1201 17:44:35.550574 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0c2b-account-create-update-gqzxq" Dec 01 17:44:36 crc kubenswrapper[4868]: I1201 17:44:36.283532 4868 generic.go:334] "Generic (PLEG): container finished" podID="168bf194-0fdb-4919-89d9-cf91e0395c21" containerID="cf1b13312b06ed5a6f43881a3f655dcda1ff81353748e85d8a3e963732db1509" exitCode=0 Dec 01 17:44:36 crc kubenswrapper[4868]: I1201 17:44:36.283586 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"168bf194-0fdb-4919-89d9-cf91e0395c21","Type":"ContainerDied","Data":"cf1b13312b06ed5a6f43881a3f655dcda1ff81353748e85d8a3e963732db1509"} Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.197839 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.295586 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-config-data\") pod \"c4416e29-710d-4683-9c40-92ca09b0e0bd\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.295673 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4416e29-710d-4683-9c40-92ca09b0e0bd-log-httpd\") pod \"c4416e29-710d-4683-9c40-92ca09b0e0bd\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.295802 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-scripts\") pod \"c4416e29-710d-4683-9c40-92ca09b0e0bd\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.295873 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-sg-core-conf-yaml\") pod \"c4416e29-710d-4683-9c40-92ca09b0e0bd\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.295899 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbxtj\" (UniqueName: \"kubernetes.io/projected/c4416e29-710d-4683-9c40-92ca09b0e0bd-kube-api-access-lbxtj\") pod \"c4416e29-710d-4683-9c40-92ca09b0e0bd\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.296065 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4416e29-710d-4683-9c40-92ca09b0e0bd-run-httpd\") pod \"c4416e29-710d-4683-9c40-92ca09b0e0bd\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.296134 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-combined-ca-bundle\") pod \"c4416e29-710d-4683-9c40-92ca09b0e0bd\" (UID: \"c4416e29-710d-4683-9c40-92ca09b0e0bd\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.297723 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4416e29-710d-4683-9c40-92ca09b0e0bd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c4416e29-710d-4683-9c40-92ca09b0e0bd" (UID: "c4416e29-710d-4683-9c40-92ca09b0e0bd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.298671 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4416e29-710d-4683-9c40-92ca09b0e0bd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c4416e29-710d-4683-9c40-92ca09b0e0bd" (UID: "c4416e29-710d-4683-9c40-92ca09b0e0bd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.304133 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4416e29-710d-4683-9c40-92ca09b0e0bd-kube-api-access-lbxtj" (OuterVolumeSpecName: "kube-api-access-lbxtj") pod "c4416e29-710d-4683-9c40-92ca09b0e0bd" (UID: "c4416e29-710d-4683-9c40-92ca09b0e0bd"). InnerVolumeSpecName "kube-api-access-lbxtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.305645 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4416e29-710d-4683-9c40-92ca09b0e0bd","Type":"ContainerDied","Data":"384485f0c86132095c26f2ac9c2f86bd5a951572999624cb434903ede448742e"} Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.305719 4868 scope.go:117] "RemoveContainer" containerID="7024addc687087043e4fb44941d29311179976a32ceaac0d20843ca1623961eb" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.305924 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.306188 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-scripts" (OuterVolumeSpecName: "scripts") pod "c4416e29-710d-4683-9c40-92ca09b0e0bd" (UID: "c4416e29-710d-4683-9c40-92ca09b0e0bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.363027 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c4416e29-710d-4683-9c40-92ca09b0e0bd" (UID: "c4416e29-710d-4683-9c40-92ca09b0e0bd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.380284 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4416e29-710d-4683-9c40-92ca09b0e0bd" (UID: "c4416e29-710d-4683-9c40-92ca09b0e0bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.403563 4868 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.403596 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbxtj\" (UniqueName: \"kubernetes.io/projected/c4416e29-710d-4683-9c40-92ca09b0e0bd-kube-api-access-lbxtj\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.403606 4868 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4416e29-710d-4683-9c40-92ca09b0e0bd-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.403616 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.403626 4868 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4416e29-710d-4683-9c40-92ca09b0e0bd-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.403633 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.415126 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-config-data" (OuterVolumeSpecName: "config-data") pod "c4416e29-710d-4683-9c40-92ca09b0e0bd" (UID: "c4416e29-710d-4683-9c40-92ca09b0e0bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.432865 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.451152 4868 scope.go:117] "RemoveContainer" containerID="800e35a04f95ccb017cb1be7f474f630a8dcd89842c365e165b16af4786ea3a5" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.505074 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-logs\") pod \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.505642 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-public-tls-certs\") pod \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.505787 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqxxg\" (UniqueName: \"kubernetes.io/projected/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-kube-api-access-bqxxg\") pod \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.505835 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-combined-ca-bundle\") pod \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.505911 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-httpd-run\") pod \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.505969 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-scripts\") pod \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.506011 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.506072 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-logs" (OuterVolumeSpecName: "logs") pod "f6e2a84b-c570-44fa-9eba-1fbe11b761eb" (UID: "f6e2a84b-c570-44fa-9eba-1fbe11b761eb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.506101 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-config-data\") pod \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\" (UID: \"f6e2a84b-c570-44fa-9eba-1fbe11b761eb\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.507273 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4416e29-710d-4683-9c40-92ca09b0e0bd-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.507299 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.510539 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f6e2a84b-c570-44fa-9eba-1fbe11b761eb" (UID: "f6e2a84b-c570-44fa-9eba-1fbe11b761eb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.523696 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-kube-api-access-bqxxg" (OuterVolumeSpecName: "kube-api-access-bqxxg") pod "f6e2a84b-c570-44fa-9eba-1fbe11b761eb" (UID: "f6e2a84b-c570-44fa-9eba-1fbe11b761eb"). InnerVolumeSpecName "kube-api-access-bqxxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.524078 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-scripts" (OuterVolumeSpecName: "scripts") pod "f6e2a84b-c570-44fa-9eba-1fbe11b761eb" (UID: "f6e2a84b-c570-44fa-9eba-1fbe11b761eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.532020 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "f6e2a84b-c570-44fa-9eba-1fbe11b761eb" (UID: "f6e2a84b-c570-44fa-9eba-1fbe11b761eb"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.545282 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xgtmd"] Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.549365 4868 scope.go:117] "RemoveContainer" containerID="e4775176a990584d72b3ad230e778ce5d004e058505e31837d7f84fd14e439f0" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.571869 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-0c2b-account-create-update-gqzxq"] Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.581426 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-xfwp6"] Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.586093 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6e2a84b-c570-44fa-9eba-1fbe11b761eb" (UID: "f6e2a84b-c570-44fa-9eba-1fbe11b761eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.594088 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f6e2a84b-c570-44fa-9eba-1fbe11b761eb" (UID: "f6e2a84b-c570-44fa-9eba-1fbe11b761eb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.598275 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-config-data" (OuterVolumeSpecName: "config-data") pod "f6e2a84b-c570-44fa-9eba-1fbe11b761eb" (UID: "f6e2a84b-c570-44fa-9eba-1fbe11b761eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.611528 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqxxg\" (UniqueName: \"kubernetes.io/projected/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-kube-api-access-bqxxg\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.611558 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.611607 4868 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.611621 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.611666 4868 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.611677 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.611685 4868 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6e2a84b-c570-44fa-9eba-1fbe11b761eb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.659657 4868 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.660098 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.717101 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/168bf194-0fdb-4919-89d9-cf91e0395c21-httpd-run\") pod \"168bf194-0fdb-4919-89d9-cf91e0395c21\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.717194 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"168bf194-0fdb-4919-89d9-cf91e0395c21\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.717308 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-scripts\") pod \"168bf194-0fdb-4919-89d9-cf91e0395c21\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.717375 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-internal-tls-certs\") pod \"168bf194-0fdb-4919-89d9-cf91e0395c21\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.717471 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjzv5\" (UniqueName: \"kubernetes.io/projected/168bf194-0fdb-4919-89d9-cf91e0395c21-kube-api-access-hjzv5\") pod \"168bf194-0fdb-4919-89d9-cf91e0395c21\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.717536 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-combined-ca-bundle\") pod \"168bf194-0fdb-4919-89d9-cf91e0395c21\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.717554 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-config-data\") pod \"168bf194-0fdb-4919-89d9-cf91e0395c21\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.717640 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/168bf194-0fdb-4919-89d9-cf91e0395c21-logs\") pod \"168bf194-0fdb-4919-89d9-cf91e0395c21\" (UID: \"168bf194-0fdb-4919-89d9-cf91e0395c21\") " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.718128 4868 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.718812 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/168bf194-0fdb-4919-89d9-cf91e0395c21-logs" (OuterVolumeSpecName: "logs") pod "168bf194-0fdb-4919-89d9-cf91e0395c21" (UID: "168bf194-0fdb-4919-89d9-cf91e0395c21"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.737426 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/168bf194-0fdb-4919-89d9-cf91e0395c21-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "168bf194-0fdb-4919-89d9-cf91e0395c21" (UID: "168bf194-0fdb-4919-89d9-cf91e0395c21"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.754982 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "168bf194-0fdb-4919-89d9-cf91e0395c21" (UID: "168bf194-0fdb-4919-89d9-cf91e0395c21"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.758152 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.771300 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.771466 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/168bf194-0fdb-4919-89d9-cf91e0395c21-kube-api-access-hjzv5" (OuterVolumeSpecName: "kube-api-access-hjzv5") pod "168bf194-0fdb-4919-89d9-cf91e0395c21" (UID: "168bf194-0fdb-4919-89d9-cf91e0395c21"). InnerVolumeSpecName "kube-api-access-hjzv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.777393 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-scripts" (OuterVolumeSpecName: "scripts") pod "168bf194-0fdb-4919-89d9-cf91e0395c21" (UID: "168bf194-0fdb-4919-89d9-cf91e0395c21"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.781444 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-wh5cc"] Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.800254 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3252-account-create-update-jrq2v"] Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.825605 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.825834 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjzv5\" (UniqueName: \"kubernetes.io/projected/168bf194-0fdb-4919-89d9-cf91e0395c21-kube-api-access-hjzv5\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.825846 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/168bf194-0fdb-4919-89d9-cf91e0395c21-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.825855 4868 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/168bf194-0fdb-4919-89d9-cf91e0395c21-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.825917 4868 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.859505 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:44:37 crc kubenswrapper[4868]: E1201 17:44:37.861308 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="168bf194-0fdb-4919-89d9-cf91e0395c21" containerName="glance-log" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.861334 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="168bf194-0fdb-4919-89d9-cf91e0395c21" containerName="glance-log" Dec 01 17:44:37 crc kubenswrapper[4868]: E1201 17:44:37.861387 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6e2a84b-c570-44fa-9eba-1fbe11b761eb" containerName="glance-log" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.861396 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6e2a84b-c570-44fa-9eba-1fbe11b761eb" containerName="glance-log" Dec 01 17:44:37 crc kubenswrapper[4868]: E1201 17:44:37.861415 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerName="proxy-httpd" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.861423 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerName="proxy-httpd" Dec 01 17:44:37 crc kubenswrapper[4868]: E1201 17:44:37.861443 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6e2a84b-c570-44fa-9eba-1fbe11b761eb" containerName="glance-httpd" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.861452 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6e2a84b-c570-44fa-9eba-1fbe11b761eb" containerName="glance-httpd" Dec 01 17:44:37 crc kubenswrapper[4868]: E1201 17:44:37.861465 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerName="sg-core" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.861634 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerName="sg-core" Dec 01 17:44:37 crc kubenswrapper[4868]: E1201 17:44:37.861659 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerName="ceilometer-notification-agent" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.861672 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerName="ceilometer-notification-agent" Dec 01 17:44:37 crc kubenswrapper[4868]: E1201 17:44:37.861704 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="168bf194-0fdb-4919-89d9-cf91e0395c21" containerName="glance-httpd" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.861711 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="168bf194-0fdb-4919-89d9-cf91e0395c21" containerName="glance-httpd" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.862978 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6e2a84b-c570-44fa-9eba-1fbe11b761eb" containerName="glance-httpd" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.863008 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="168bf194-0fdb-4919-89d9-cf91e0395c21" containerName="glance-log" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.863277 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="168bf194-0fdb-4919-89d9-cf91e0395c21" containerName="glance-httpd" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.863328 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerName="ceilometer-notification-agent" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.863347 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerName="proxy-httpd" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.863379 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6e2a84b-c570-44fa-9eba-1fbe11b761eb" containerName="glance-log" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.863396 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4416e29-710d-4683-9c40-92ca09b0e0bd" containerName="sg-core" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.870393 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.874280 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.874857 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.885901 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-bc8a-account-create-update-ht62v"] Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.906647 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.934440 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-config-data\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.934590 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-scripts\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.934667 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqhg4\" (UniqueName: \"kubernetes.io/projected/ef3afc2e-b496-4104-98cf-35507acfe031-kube-api-access-hqhg4\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.934822 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef3afc2e-b496-4104-98cf-35507acfe031-run-httpd\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.934910 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.935015 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.935102 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef3afc2e-b496-4104-98cf-35507acfe031-log-httpd\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.939968 4868 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.969705 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "168bf194-0fdb-4919-89d9-cf91e0395c21" (UID: "168bf194-0fdb-4919-89d9-cf91e0395c21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.992353 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-config-data" (OuterVolumeSpecName: "config-data") pod "168bf194-0fdb-4919-89d9-cf91e0395c21" (UID: "168bf194-0fdb-4919-89d9-cf91e0395c21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:37 crc kubenswrapper[4868]: I1201 17:44:37.995178 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "168bf194-0fdb-4919-89d9-cf91e0395c21" (UID: "168bf194-0fdb-4919-89d9-cf91e0395c21"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.038428 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-scripts\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.038474 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqhg4\" (UniqueName: \"kubernetes.io/projected/ef3afc2e-b496-4104-98cf-35507acfe031-kube-api-access-hqhg4\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.038529 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef3afc2e-b496-4104-98cf-35507acfe031-run-httpd\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.038571 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.038598 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.038632 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef3afc2e-b496-4104-98cf-35507acfe031-log-httpd\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.038934 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-config-data\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.039017 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.039030 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.039041 4868 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.039050 4868 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/168bf194-0fdb-4919-89d9-cf91e0395c21-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.039975 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef3afc2e-b496-4104-98cf-35507acfe031-run-httpd\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.042040 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef3afc2e-b496-4104-98cf-35507acfe031-log-httpd\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.051162 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.055970 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-config-data\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.056110 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.061474 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqhg4\" (UniqueName: \"kubernetes.io/projected/ef3afc2e-b496-4104-98cf-35507acfe031-kube-api-access-hqhg4\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.074312 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-scripts\") pod \"ceilometer-0\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " pod="openstack/ceilometer-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.204681 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4416e29-710d-4683-9c40-92ca09b0e0bd" path="/var/lib/kubelet/pods/c4416e29-710d-4683-9c40-92ca09b0e0bd/volumes" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.327452 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-bc8a-account-create-update-ht62v" event={"ID":"a996536c-2c87-4c80-90d6-c6758c109af1","Type":"ContainerStarted","Data":"d8be20d17f15cdeb28eb41fc65880efb9af9883fd88674014078a4f02bc7ffcc"} Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.329408 4868 generic.go:334] "Generic (PLEG): container finished" podID="31d8af7d-f71e-42be-ab59-237f2d505b1f" containerID="8b0f1c867c810bfa762034c066053aa1e68fc4f3e7063c956aec64f2ef3f0c20" exitCode=0 Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.329471 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xgtmd" event={"ID":"31d8af7d-f71e-42be-ab59-237f2d505b1f","Type":"ContainerDied","Data":"8b0f1c867c810bfa762034c066053aa1e68fc4f3e7063c956aec64f2ef3f0c20"} Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.329491 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xgtmd" event={"ID":"31d8af7d-f71e-42be-ab59-237f2d505b1f","Type":"ContainerStarted","Data":"54ecb85c3a556be293d6b1d9bf3314d2e79c63264c7d4e22ae930f812c1828e9"} Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.335956 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3f3f5b1c-3ed5-4d79-b298-474f483b2c4a","Type":"ContainerStarted","Data":"d25807ce555a382c745c819d7285fcf184a93cfbb99eedea74b5d9e99bb2c08f"} Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.340235 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wh5cc" event={"ID":"288e3235-6690-4c67-9e4f-2735be76d249","Type":"ContainerStarted","Data":"3675bf5cb00703490e398ade7bf7cab6b19f97a4f39b765c1382c0d775abf7a3"} Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.342588 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xfwp6" event={"ID":"b801bf8a-a39f-4537-aa83-06e958c31e02","Type":"ContainerStarted","Data":"fbe62dc81d0ad9dbb416e41716f138a5c946c5655a34c6683a1cda441c58faae"} Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.347468 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f6e2a84b-c570-44fa-9eba-1fbe11b761eb","Type":"ContainerDied","Data":"1744ea545b02cbc388fac71ddd1ac84901c36a0c2140e2e14b3dc263e4cab5ca"} Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.347974 4868 scope.go:117] "RemoveContainer" containerID="277c8c1ea3ef347a4fac8ddbe7c0add43671c594e88b66a7eee55784ddd140eb" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.347654 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.351745 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3252-account-create-update-jrq2v" event={"ID":"e5951a13-5101-4956-bcf6-345fab68ea75","Type":"ContainerStarted","Data":"f4087b89ee4fabdcef2a8999b88f8a7af360bc1b24464dfb83796be60067f798"} Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.361545 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:44:38 crc kubenswrapper[4868]: E1201 17:44:38.362230 4868 kuberuntime_gc.go:389] "Failed to remove container log dead symlink" err="remove /var/log/containers/glance-default-external-api-0_openstack_glance-httpd-277c8c1ea3ef347a4fac8ddbe7c0add43671c594e88b66a7eee55784ddd140eb.log: no such file or directory" path="/var/log/containers/glance-default-external-api-0_openstack_glance-httpd-277c8c1ea3ef347a4fac8ddbe7c0add43671c594e88b66a7eee55784ddd140eb.log" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.364668 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"168bf194-0fdb-4919-89d9-cf91e0395c21","Type":"ContainerDied","Data":"bfe6500751e8289183c54f560ff0195b530c0855a2dcef8214c466a35995ce58"} Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.365085 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.370644 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0c2b-account-create-update-gqzxq" event={"ID":"441fc0fe-f3fe-4e94-abd8-0d126b289042","Type":"ContainerStarted","Data":"fc0f426efbec3966f324d8902896453541e102de252ff24ad7f55198dfdad0aa"} Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.393418 4868 scope.go:117] "RemoveContainer" containerID="b9832a951b9d366bd17c83326764a4e46db1979d6f50477ef7fd5e2ea671ac9d" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.402518 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.752535859 podStartE2EDuration="13.402495361s" podCreationTimestamp="2025-12-01 17:44:25 +0000 UTC" firstStartedPulling="2025-12-01 17:44:26.13282169 +0000 UTC m=+1138.503932101" lastFinishedPulling="2025-12-01 17:44:36.782781192 +0000 UTC m=+1149.153891603" observedRunningTime="2025-12-01 17:44:38.378904165 +0000 UTC m=+1150.750014576" watchObservedRunningTime="2025-12-01 17:44:38.402495361 +0000 UTC m=+1150.773605772" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.450994 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.500201 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.519543 4868 scope.go:117] "RemoveContainer" containerID="cf1b13312b06ed5a6f43881a3f655dcda1ff81353748e85d8a3e963732db1509" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.526585 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.544344 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.546447 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.551462 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.551676 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5tv8n" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.551805 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.551976 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.564315 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.575711 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.583936 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.585563 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.588062 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.588316 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.593786 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.654459 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.654507 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.654529 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4247f042-96f0-4bc5-9ca3-a64ee5616624-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.654555 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.654581 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-logs\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.654595 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.654610 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4247f042-96f0-4bc5-9ca3-a64ee5616624-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.654638 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx7vs\" (UniqueName: \"kubernetes.io/projected/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-kube-api-access-qx7vs\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.654665 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4247f042-96f0-4bc5-9ca3-a64ee5616624-scripts\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.654683 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.654714 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnv2q\" (UniqueName: \"kubernetes.io/projected/4247f042-96f0-4bc5-9ca3-a64ee5616624-kube-api-access-fnv2q\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.654733 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4247f042-96f0-4bc5-9ca3-a64ee5616624-config-data\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.654768 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4247f042-96f0-4bc5-9ca3-a64ee5616624-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.654803 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.655004 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.655036 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4247f042-96f0-4bc5-9ca3-a64ee5616624-logs\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.670933 4868 scope.go:117] "RemoveContainer" containerID="5dd285127895f6cb2cb5b9ca97f148b3d95e701f72c41175af3d807c0a595a59" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.761507 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4247f042-96f0-4bc5-9ca3-a64ee5616624-logs\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.762004 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.762043 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.762072 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.762570 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4247f042-96f0-4bc5-9ca3-a64ee5616624-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.762599 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.762996 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-logs\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.763033 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4247f042-96f0-4bc5-9ca3-a64ee5616624-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.763054 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.763087 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx7vs\" (UniqueName: \"kubernetes.io/projected/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-kube-api-access-qx7vs\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.763119 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4247f042-96f0-4bc5-9ca3-a64ee5616624-scripts\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.763138 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.763174 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnv2q\" (UniqueName: \"kubernetes.io/projected/4247f042-96f0-4bc5-9ca3-a64ee5616624-kube-api-access-fnv2q\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.763195 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4247f042-96f0-4bc5-9ca3-a64ee5616624-config-data\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.763359 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4247f042-96f0-4bc5-9ca3-a64ee5616624-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.763392 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.763964 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.765150 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4247f042-96f0-4bc5-9ca3-a64ee5616624-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.765407 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.765692 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4247f042-96f0-4bc5-9ca3-a64ee5616624-logs\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.766736 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-logs\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.767083 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.784988 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.797113 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.797612 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4247f042-96f0-4bc5-9ca3-a64ee5616624-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.799776 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.803234 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4247f042-96f0-4bc5-9ca3-a64ee5616624-scripts\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.804458 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx7vs\" (UniqueName: \"kubernetes.io/projected/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-kube-api-access-qx7vs\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.804646 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4247f042-96f0-4bc5-9ca3-a64ee5616624-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.805300 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4247f042-96f0-4bc5-9ca3-a64ee5616624-config-data\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.844798 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnv2q\" (UniqueName: \"kubernetes.io/projected/4247f042-96f0-4bc5-9ca3-a64ee5616624-kube-api-access-fnv2q\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.913522 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bc0d7b8-d497-4339-ad24-c9b75e6e06fc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.917989 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"4247f042-96f0-4bc5-9ca3-a64ee5616624\") " pod="openstack/glance-default-external-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.973733 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc\") " pod="openstack/glance-default-internal-api-0" Dec 01 17:44:38 crc kubenswrapper[4868]: I1201 17:44:38.984349 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 17:44:39 crc kubenswrapper[4868]: E1201 17:44:39.104268 4868 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/315ec96b112d8675d55e535602c17394654ddedfeb1278f7d851dfaa6f508b69/diff" to get inode usage: stat /var/lib/containers/storage/overlay/315ec96b112d8675d55e535602c17394654ddedfeb1278f7d851dfaa6f508b69/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_glance-default-external-api-0_f6e2a84b-c570-44fa-9eba-1fbe11b761eb/glance-httpd/0.log" to get inode usage: stat /var/log/pods/openstack_glance-default-external-api-0_f6e2a84b-c570-44fa-9eba-1fbe11b761eb/glance-httpd/0.log: no such file or directory Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.169713 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.208409 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.210079 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-86984fb57-t5hzr" Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.273259 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.391558 4868 generic.go:334] "Generic (PLEG): container finished" podID="b801bf8a-a39f-4537-aa83-06e958c31e02" containerID="17e85ef8e3bd3ce5537753a9525c2ea00ed57ec7e412ffc6e943fddc8d20a79e" exitCode=0 Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.391657 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xfwp6" event={"ID":"b801bf8a-a39f-4537-aa83-06e958c31e02","Type":"ContainerDied","Data":"17e85ef8e3bd3ce5537753a9525c2ea00ed57ec7e412ffc6e943fddc8d20a79e"} Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.394031 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3252-account-create-update-jrq2v" event={"ID":"e5951a13-5101-4956-bcf6-345fab68ea75","Type":"ContainerDied","Data":"7d357e0acb6033a19187106e62ec3aa8e00109ee4c4372d7d5c56a01923a0553"} Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.394630 4868 generic.go:334] "Generic (PLEG): container finished" podID="e5951a13-5101-4956-bcf6-345fab68ea75" containerID="7d357e0acb6033a19187106e62ec3aa8e00109ee4c4372d7d5c56a01923a0553" exitCode=0 Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.401731 4868 generic.go:334] "Generic (PLEG): container finished" podID="441fc0fe-f3fe-4e94-abd8-0d126b289042" containerID="418d436c4bbfe6f81c8b65515afb9a59fc9bd4230c8346a97475719320ad3d9a" exitCode=0 Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.401851 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0c2b-account-create-update-gqzxq" event={"ID":"441fc0fe-f3fe-4e94-abd8-0d126b289042","Type":"ContainerDied","Data":"418d436c4bbfe6f81c8b65515afb9a59fc9bd4230c8346a97475719320ad3d9a"} Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.403312 4868 generic.go:334] "Generic (PLEG): container finished" podID="a996536c-2c87-4c80-90d6-c6758c109af1" containerID="0878faec332107965aaf8ad00d9fdf2258d16edd2098a54733a1bc3c889c111f" exitCode=0 Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.403387 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-bc8a-account-create-update-ht62v" event={"ID":"a996536c-2c87-4c80-90d6-c6758c109af1","Type":"ContainerDied","Data":"0878faec332107965aaf8ad00d9fdf2258d16edd2098a54733a1bc3c889c111f"} Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.404856 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef3afc2e-b496-4104-98cf-35507acfe031","Type":"ContainerStarted","Data":"f4627629b250f4a68ebef59261453e39d521b7684c91a1f883c2509678999b98"} Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.406310 4868 generic.go:334] "Generic (PLEG): container finished" podID="288e3235-6690-4c67-9e4f-2735be76d249" containerID="80b8fe19cab38a4094b4a12f3720205d98255899f44e9c3b71dc6387a57e148d" exitCode=0 Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.406374 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wh5cc" event={"ID":"288e3235-6690-4c67-9e4f-2735be76d249","Type":"ContainerDied","Data":"80b8fe19cab38a4094b4a12f3720205d98255899f44e9c3b71dc6387a57e148d"} Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.690850 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.850813 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xgtmd" Dec 01 17:44:39 crc kubenswrapper[4868]: I1201 17:44:39.954915 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.009004 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31d8af7d-f71e-42be-ab59-237f2d505b1f-operator-scripts\") pod \"31d8af7d-f71e-42be-ab59-237f2d505b1f\" (UID: \"31d8af7d-f71e-42be-ab59-237f2d505b1f\") " Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.009365 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5vjj\" (UniqueName: \"kubernetes.io/projected/31d8af7d-f71e-42be-ab59-237f2d505b1f-kube-api-access-k5vjj\") pod \"31d8af7d-f71e-42be-ab59-237f2d505b1f\" (UID: \"31d8af7d-f71e-42be-ab59-237f2d505b1f\") " Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.011109 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8af7d-f71e-42be-ab59-237f2d505b1f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "31d8af7d-f71e-42be-ab59-237f2d505b1f" (UID: "31d8af7d-f71e-42be-ab59-237f2d505b1f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.016351 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8af7d-f71e-42be-ab59-237f2d505b1f-kube-api-access-k5vjj" (OuterVolumeSpecName: "kube-api-access-k5vjj") pod "31d8af7d-f71e-42be-ab59-237f2d505b1f" (UID: "31d8af7d-f71e-42be-ab59-237f2d505b1f"). InnerVolumeSpecName "kube-api-access-k5vjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.112349 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31d8af7d-f71e-42be-ab59-237f2d505b1f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.112384 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5vjj\" (UniqueName: \"kubernetes.io/projected/31d8af7d-f71e-42be-ab59-237f2d505b1f-kube-api-access-k5vjj\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.194096 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="168bf194-0fdb-4919-89d9-cf91e0395c21" path="/var/lib/kubelet/pods/168bf194-0fdb-4919-89d9-cf91e0395c21/volumes" Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.194993 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6e2a84b-c570-44fa-9eba-1fbe11b761eb" path="/var/lib/kubelet/pods/f6e2a84b-c570-44fa-9eba-1fbe11b761eb/volumes" Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.437775 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef3afc2e-b496-4104-98cf-35507acfe031","Type":"ContainerStarted","Data":"3861fbd918cc8846b4d75999efd64c75f715a2f8ff83a10f1b6bff5fde69bb4e"} Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.441442 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4247f042-96f0-4bc5-9ca3-a64ee5616624","Type":"ContainerStarted","Data":"b2d1fd56ee2be98db885591c11098e62e5b89f9ac5448a24291845247efff802"} Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.444232 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xgtmd" event={"ID":"31d8af7d-f71e-42be-ab59-237f2d505b1f","Type":"ContainerDied","Data":"54ecb85c3a556be293d6b1d9bf3314d2e79c63264c7d4e22ae930f812c1828e9"} Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.444300 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54ecb85c3a556be293d6b1d9bf3314d2e79c63264c7d4e22ae930f812c1828e9" Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.444394 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xgtmd" Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.447824 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc","Type":"ContainerStarted","Data":"f5fc21c488cea64e0355d9ccfa3ac3e45cbc6e3ea3fec7d3693bf511dabf66de"} Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.795400 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3252-account-create-update-jrq2v" Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.942318 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5951a13-5101-4956-bcf6-345fab68ea75-operator-scripts\") pod \"e5951a13-5101-4956-bcf6-345fab68ea75\" (UID: \"e5951a13-5101-4956-bcf6-345fab68ea75\") " Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.942997 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzcdr\" (UniqueName: \"kubernetes.io/projected/e5951a13-5101-4956-bcf6-345fab68ea75-kube-api-access-vzcdr\") pod \"e5951a13-5101-4956-bcf6-345fab68ea75\" (UID: \"e5951a13-5101-4956-bcf6-345fab68ea75\") " Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.942959 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5951a13-5101-4956-bcf6-345fab68ea75-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5951a13-5101-4956-bcf6-345fab68ea75" (UID: "e5951a13-5101-4956-bcf6-345fab68ea75"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:40 crc kubenswrapper[4868]: I1201 17:44:40.989144 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5951a13-5101-4956-bcf6-345fab68ea75-kube-api-access-vzcdr" (OuterVolumeSpecName: "kube-api-access-vzcdr") pod "e5951a13-5101-4956-bcf6-345fab68ea75" (UID: "e5951a13-5101-4956-bcf6-345fab68ea75"). InnerVolumeSpecName "kube-api-access-vzcdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.054033 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5951a13-5101-4956-bcf6-345fab68ea75-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.054358 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzcdr\" (UniqueName: \"kubernetes.io/projected/e5951a13-5101-4956-bcf6-345fab68ea75-kube-api-access-vzcdr\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: W1201 17:44:41.100428 4868 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31d8af7d_f71e_42be_ab59_237f2d505b1f.slice/crio-54ecb85c3a556be293d6b1d9bf3314d2e79c63264c7d4e22ae930f812c1828e9": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31d8af7d_f71e_42be_ab59_237f2d505b1f.slice/crio-54ecb85c3a556be293d6b1d9bf3314d2e79c63264c7d4e22ae930f812c1828e9: no such file or directory Dec 01 17:44:41 crc kubenswrapper[4868]: W1201 17:44:41.101541 4868 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31d8af7d_f71e_42be_ab59_237f2d505b1f.slice/crio-conmon-8b0f1c867c810bfa762034c066053aa1e68fc4f3e7063c956aec64f2ef3f0c20.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31d8af7d_f71e_42be_ab59_237f2d505b1f.slice/crio-conmon-8b0f1c867c810bfa762034c066053aa1e68fc4f3e7063c956aec64f2ef3f0c20.scope: no such file or directory Dec 01 17:44:41 crc kubenswrapper[4868]: W1201 17:44:41.101572 4868 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31d8af7d_f71e_42be_ab59_237f2d505b1f.slice/crio-8b0f1c867c810bfa762034c066053aa1e68fc4f3e7063c956aec64f2ef3f0c20.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31d8af7d_f71e_42be_ab59_237f2d505b1f.slice/crio-8b0f1c867c810bfa762034c066053aa1e68fc4f3e7063c956aec64f2ef3f0c20.scope: no such file or directory Dec 01 17:44:41 crc kubenswrapper[4868]: W1201 17:44:41.101703 4868 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb801bf8a_a39f_4537_aa83_06e958c31e02.slice/crio-conmon-17e85ef8e3bd3ce5537753a9525c2ea00ed57ec7e412ffc6e943fddc8d20a79e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb801bf8a_a39f_4537_aa83_06e958c31e02.slice/crio-conmon-17e85ef8e3bd3ce5537753a9525c2ea00ed57ec7e412ffc6e943fddc8d20a79e.scope: no such file or directory Dec 01 17:44:41 crc kubenswrapper[4868]: W1201 17:44:41.101738 4868 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5951a13_5101_4956_bcf6_345fab68ea75.slice/crio-conmon-7d357e0acb6033a19187106e62ec3aa8e00109ee4c4372d7d5c56a01923a0553.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5951a13_5101_4956_bcf6_345fab68ea75.slice/crio-conmon-7d357e0acb6033a19187106e62ec3aa8e00109ee4c4372d7d5c56a01923a0553.scope: no such file or directory Dec 01 17:44:41 crc kubenswrapper[4868]: W1201 17:44:41.101785 4868 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda996536c_2c87_4c80_90d6_c6758c109af1.slice/crio-conmon-0878faec332107965aaf8ad00d9fdf2258d16edd2098a54733a1bc3c889c111f.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda996536c_2c87_4c80_90d6_c6758c109af1.slice/crio-conmon-0878faec332107965aaf8ad00d9fdf2258d16edd2098a54733a1bc3c889c111f.scope: no such file or directory Dec 01 17:44:41 crc kubenswrapper[4868]: W1201 17:44:41.106816 4868 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda996536c_2c87_4c80_90d6_c6758c109af1.slice/crio-0878faec332107965aaf8ad00d9fdf2258d16edd2098a54733a1bc3c889c111f.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda996536c_2c87_4c80_90d6_c6758c109af1.slice/crio-0878faec332107965aaf8ad00d9fdf2258d16edd2098a54733a1bc3c889c111f.scope: no such file or directory Dec 01 17:44:41 crc kubenswrapper[4868]: W1201 17:44:41.106861 4868 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb801bf8a_a39f_4537_aa83_06e958c31e02.slice/crio-17e85ef8e3bd3ce5537753a9525c2ea00ed57ec7e412ffc6e943fddc8d20a79e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb801bf8a_a39f_4537_aa83_06e958c31e02.slice/crio-17e85ef8e3bd3ce5537753a9525c2ea00ed57ec7e412ffc6e943fddc8d20a79e.scope: no such file or directory Dec 01 17:44:41 crc kubenswrapper[4868]: W1201 17:44:41.106880 4868 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5951a13_5101_4956_bcf6_345fab68ea75.slice/crio-7d357e0acb6033a19187106e62ec3aa8e00109ee4c4372d7d5c56a01923a0553.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5951a13_5101_4956_bcf6_345fab68ea75.slice/crio-7d357e0acb6033a19187106e62ec3aa8e00109ee4c4372d7d5c56a01923a0553.scope: no such file or directory Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.314634 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xfwp6" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.357711 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bc8a-account-create-update-ht62v" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.397288 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wh5cc" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.398138 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0c2b-account-create-update-gqzxq" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.445557 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.469401 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a996536c-2c87-4c80-90d6-c6758c109af1-operator-scripts\") pod \"a996536c-2c87-4c80-90d6-c6758c109af1\" (UID: \"a996536c-2c87-4c80-90d6-c6758c109af1\") " Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.469465 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b801bf8a-a39f-4537-aa83-06e958c31e02-operator-scripts\") pod \"b801bf8a-a39f-4537-aa83-06e958c31e02\" (UID: \"b801bf8a-a39f-4537-aa83-06e958c31e02\") " Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.469631 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6m29\" (UniqueName: \"kubernetes.io/projected/a996536c-2c87-4c80-90d6-c6758c109af1-kube-api-access-d6m29\") pod \"a996536c-2c87-4c80-90d6-c6758c109af1\" (UID: \"a996536c-2c87-4c80-90d6-c6758c109af1\") " Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.469895 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kq7k\" (UniqueName: \"kubernetes.io/projected/b801bf8a-a39f-4537-aa83-06e958c31e02-kube-api-access-8kq7k\") pod \"b801bf8a-a39f-4537-aa83-06e958c31e02\" (UID: \"b801bf8a-a39f-4537-aa83-06e958c31e02\") " Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.474917 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b801bf8a-a39f-4537-aa83-06e958c31e02-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b801bf8a-a39f-4537-aa83-06e958c31e02" (UID: "b801bf8a-a39f-4537-aa83-06e958c31e02"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.475646 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a996536c-2c87-4c80-90d6-c6758c109af1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a996536c-2c87-4c80-90d6-c6758c109af1" (UID: "a996536c-2c87-4c80-90d6-c6758c109af1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.494305 4868 generic.go:334] "Generic (PLEG): container finished" podID="fdb719ec-3812-4252-bdb9-1bc93aa35462" containerID="98299ac37c9d398981045965e1ed7484fda45b7142252554239df2eab9ac0ba1" exitCode=137 Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.494453 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-97868dbf6-djbqg" event={"ID":"fdb719ec-3812-4252-bdb9-1bc93aa35462","Type":"ContainerDied","Data":"98299ac37c9d398981045965e1ed7484fda45b7142252554239df2eab9ac0ba1"} Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.494485 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-97868dbf6-djbqg" event={"ID":"fdb719ec-3812-4252-bdb9-1bc93aa35462","Type":"ContainerDied","Data":"795d41c1d61d507152f36a04c864649d2d3fb86f81c66835fb220bb1ba93df8d"} Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.494521 4868 scope.go:117] "RemoveContainer" containerID="5dbf91a124b9ea9b19e7613ba78db4552ecf9b204c056b4966d3fd691eb80b32" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.494641 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b801bf8a-a39f-4537-aa83-06e958c31e02-kube-api-access-8kq7k" (OuterVolumeSpecName: "kube-api-access-8kq7k") pod "b801bf8a-a39f-4537-aa83-06e958c31e02" (UID: "b801bf8a-a39f-4537-aa83-06e958c31e02"). InnerVolumeSpecName "kube-api-access-8kq7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.494720 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-97868dbf6-djbqg" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.494920 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a996536c-2c87-4c80-90d6-c6758c109af1-kube-api-access-d6m29" (OuterVolumeSpecName: "kube-api-access-d6m29") pod "a996536c-2c87-4c80-90d6-c6758c109af1" (UID: "a996536c-2c87-4c80-90d6-c6758c109af1"). InnerVolumeSpecName "kube-api-access-d6m29". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.503957 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-bc8a-account-create-update-ht62v" event={"ID":"a996536c-2c87-4c80-90d6-c6758c109af1","Type":"ContainerDied","Data":"d8be20d17f15cdeb28eb41fc65880efb9af9883fd88674014078a4f02bc7ffcc"} Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.504227 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8be20d17f15cdeb28eb41fc65880efb9af9883fd88674014078a4f02bc7ffcc" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.504374 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bc8a-account-create-update-ht62v" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.512474 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc","Type":"ContainerStarted","Data":"8d748b1fbca8c4b85271111b11cc1714867f8d0c055159b75f0950f27018e397"} Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.514867 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0c2b-account-create-update-gqzxq" event={"ID":"441fc0fe-f3fe-4e94-abd8-0d126b289042","Type":"ContainerDied","Data":"fc0f426efbec3966f324d8902896453541e102de252ff24ad7f55198dfdad0aa"} Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.514900 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc0f426efbec3966f324d8902896453541e102de252ff24ad7f55198dfdad0aa" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.514978 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0c2b-account-create-update-gqzxq" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.524397 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xfwp6" event={"ID":"b801bf8a-a39f-4537-aa83-06e958c31e02","Type":"ContainerDied","Data":"fbe62dc81d0ad9dbb416e41716f138a5c946c5655a34c6683a1cda441c58faae"} Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.524447 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbe62dc81d0ad9dbb416e41716f138a5c946c5655a34c6683a1cda441c58faae" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.524521 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xfwp6" Dec 01 17:44:41 crc kubenswrapper[4868]: E1201 17:44:41.529855 4868 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdb719ec_3812_4252_bdb9_1bc93aa35462.slice/crio-conmon-98299ac37c9d398981045965e1ed7484fda45b7142252554239df2eab9ac0ba1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6e2a84b_c570_44fa_9eba_1fbe11b761eb.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6e2a84b_c570_44fa_9eba_1fbe11b761eb.slice/crio-1744ea545b02cbc388fac71ddd1ac84901c36a0c2140e2e14b3dc263e4cab5ca\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod168bf194_0fdb_4919_89d9_cf91e0395c21.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod168bf194_0fdb_4919_89d9_cf91e0395c21.slice/crio-bfe6500751e8289183c54f560ff0195b530c0855a2dcef8214c466a35995ce58\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdb719ec_3812_4252_bdb9_1bc93aa35462.slice/crio-98299ac37c9d398981045965e1ed7484fda45b7142252554239df2eab9ac0ba1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31d8af7d_f71e_42be_ab59_237f2d505b1f.slice\": RecentStats: unable to find data in memory cache]" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.531192 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3252-account-create-update-jrq2v" event={"ID":"e5951a13-5101-4956-bcf6-345fab68ea75","Type":"ContainerDied","Data":"f4087b89ee4fabdcef2a8999b88f8a7af360bc1b24464dfb83796be60067f798"} Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.531254 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4087b89ee4fabdcef2a8999b88f8a7af360bc1b24464dfb83796be60067f798" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.531476 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3252-account-create-update-jrq2v" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.538067 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef3afc2e-b496-4104-98cf-35507acfe031","Type":"ContainerStarted","Data":"9be42b8f2e80c8df4a4ff4b89367f3467b0884b1e51b787b55055283d2ba8862"} Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.560013 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4247f042-96f0-4bc5-9ca3-a64ee5616624","Type":"ContainerStarted","Data":"caeb76bc401bf03d3bd461825aa27cc3f3108b5222029a1ae67710c51253b1ef"} Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.570392 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wh5cc" event={"ID":"288e3235-6690-4c67-9e4f-2735be76d249","Type":"ContainerDied","Data":"3675bf5cb00703490e398ade7bf7cab6b19f97a4f39b765c1382c0d775abf7a3"} Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.570438 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3675bf5cb00703490e398ade7bf7cab6b19f97a4f39b765c1382c0d775abf7a3" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.570497 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wh5cc" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.571636 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7kbw\" (UniqueName: \"kubernetes.io/projected/fdb719ec-3812-4252-bdb9-1bc93aa35462-kube-api-access-f7kbw\") pod \"fdb719ec-3812-4252-bdb9-1bc93aa35462\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.571691 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-combined-ca-bundle\") pod \"fdb719ec-3812-4252-bdb9-1bc93aa35462\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.571737 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-horizon-tls-certs\") pod \"fdb719ec-3812-4252-bdb9-1bc93aa35462\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.571821 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-horizon-secret-key\") pod \"fdb719ec-3812-4252-bdb9-1bc93aa35462\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.571875 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fknrm\" (UniqueName: \"kubernetes.io/projected/288e3235-6690-4c67-9e4f-2735be76d249-kube-api-access-fknrm\") pod \"288e3235-6690-4c67-9e4f-2735be76d249\" (UID: \"288e3235-6690-4c67-9e4f-2735be76d249\") " Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.571924 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/288e3235-6690-4c67-9e4f-2735be76d249-operator-scripts\") pod \"288e3235-6690-4c67-9e4f-2735be76d249\" (UID: \"288e3235-6690-4c67-9e4f-2735be76d249\") " Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.571973 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvnfv\" (UniqueName: \"kubernetes.io/projected/441fc0fe-f3fe-4e94-abd8-0d126b289042-kube-api-access-dvnfv\") pod \"441fc0fe-f3fe-4e94-abd8-0d126b289042\" (UID: \"441fc0fe-f3fe-4e94-abd8-0d126b289042\") " Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.572017 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fdb719ec-3812-4252-bdb9-1bc93aa35462-config-data\") pod \"fdb719ec-3812-4252-bdb9-1bc93aa35462\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.572101 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/441fc0fe-f3fe-4e94-abd8-0d126b289042-operator-scripts\") pod \"441fc0fe-f3fe-4e94-abd8-0d126b289042\" (UID: \"441fc0fe-f3fe-4e94-abd8-0d126b289042\") " Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.572145 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fdb719ec-3812-4252-bdb9-1bc93aa35462-logs\") pod \"fdb719ec-3812-4252-bdb9-1bc93aa35462\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.572197 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb719ec-3812-4252-bdb9-1bc93aa35462-scripts\") pod \"fdb719ec-3812-4252-bdb9-1bc93aa35462\" (UID: \"fdb719ec-3812-4252-bdb9-1bc93aa35462\") " Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.572645 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kq7k\" (UniqueName: \"kubernetes.io/projected/b801bf8a-a39f-4537-aa83-06e958c31e02-kube-api-access-8kq7k\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.572668 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a996536c-2c87-4c80-90d6-c6758c109af1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.572682 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b801bf8a-a39f-4537-aa83-06e958c31e02-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.572695 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6m29\" (UniqueName: \"kubernetes.io/projected/a996536c-2c87-4c80-90d6-c6758c109af1-kube-api-access-d6m29\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.574380 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/288e3235-6690-4c67-9e4f-2735be76d249-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "288e3235-6690-4c67-9e4f-2735be76d249" (UID: "288e3235-6690-4c67-9e4f-2735be76d249"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.575770 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/441fc0fe-f3fe-4e94-abd8-0d126b289042-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "441fc0fe-f3fe-4e94-abd8-0d126b289042" (UID: "441fc0fe-f3fe-4e94-abd8-0d126b289042"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.576205 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdb719ec-3812-4252-bdb9-1bc93aa35462-logs" (OuterVolumeSpecName: "logs") pod "fdb719ec-3812-4252-bdb9-1bc93aa35462" (UID: "fdb719ec-3812-4252-bdb9-1bc93aa35462"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.579382 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/288e3235-6690-4c67-9e4f-2735be76d249-kube-api-access-fknrm" (OuterVolumeSpecName: "kube-api-access-fknrm") pod "288e3235-6690-4c67-9e4f-2735be76d249" (UID: "288e3235-6690-4c67-9e4f-2735be76d249"). InnerVolumeSpecName "kube-api-access-fknrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.581642 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "fdb719ec-3812-4252-bdb9-1bc93aa35462" (UID: "fdb719ec-3812-4252-bdb9-1bc93aa35462"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.582434 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdb719ec-3812-4252-bdb9-1bc93aa35462-kube-api-access-f7kbw" (OuterVolumeSpecName: "kube-api-access-f7kbw") pod "fdb719ec-3812-4252-bdb9-1bc93aa35462" (UID: "fdb719ec-3812-4252-bdb9-1bc93aa35462"). InnerVolumeSpecName "kube-api-access-f7kbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.605704 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/441fc0fe-f3fe-4e94-abd8-0d126b289042-kube-api-access-dvnfv" (OuterVolumeSpecName: "kube-api-access-dvnfv") pod "441fc0fe-f3fe-4e94-abd8-0d126b289042" (UID: "441fc0fe-f3fe-4e94-abd8-0d126b289042"). InnerVolumeSpecName "kube-api-access-dvnfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.609956 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.609911673 podStartE2EDuration="3.609911673s" podCreationTimestamp="2025-12-01 17:44:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:44:41.60614693 +0000 UTC m=+1153.977257341" watchObservedRunningTime="2025-12-01 17:44:41.609911673 +0000 UTC m=+1153.981022084" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.621065 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdb719ec-3812-4252-bdb9-1bc93aa35462-scripts" (OuterVolumeSpecName: "scripts") pod "fdb719ec-3812-4252-bdb9-1bc93aa35462" (UID: "fdb719ec-3812-4252-bdb9-1bc93aa35462"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.622261 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdb719ec-3812-4252-bdb9-1bc93aa35462-config-data" (OuterVolumeSpecName: "config-data") pod "fdb719ec-3812-4252-bdb9-1bc93aa35462" (UID: "fdb719ec-3812-4252-bdb9-1bc93aa35462"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.629829 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fdb719ec-3812-4252-bdb9-1bc93aa35462" (UID: "fdb719ec-3812-4252-bdb9-1bc93aa35462"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.674323 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7kbw\" (UniqueName: \"kubernetes.io/projected/fdb719ec-3812-4252-bdb9-1bc93aa35462-kube-api-access-f7kbw\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.674386 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.674396 4868 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.674407 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fknrm\" (UniqueName: \"kubernetes.io/projected/288e3235-6690-4c67-9e4f-2735be76d249-kube-api-access-fknrm\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.674417 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/288e3235-6690-4c67-9e4f-2735be76d249-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.674426 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvnfv\" (UniqueName: \"kubernetes.io/projected/441fc0fe-f3fe-4e94-abd8-0d126b289042-kube-api-access-dvnfv\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.674437 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fdb719ec-3812-4252-bdb9-1bc93aa35462-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.674450 4868 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/441fc0fe-f3fe-4e94-abd8-0d126b289042-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.674459 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fdb719ec-3812-4252-bdb9-1bc93aa35462-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.674469 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb719ec-3812-4252-bdb9-1bc93aa35462-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.696833 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "fdb719ec-3812-4252-bdb9-1bc93aa35462" (UID: "fdb719ec-3812-4252-bdb9-1bc93aa35462"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.743967 4868 scope.go:117] "RemoveContainer" containerID="98299ac37c9d398981045965e1ed7484fda45b7142252554239df2eab9ac0ba1" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.775780 4868 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdb719ec-3812-4252-bdb9-1bc93aa35462-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.776784 4868 scope.go:117] "RemoveContainer" containerID="5dbf91a124b9ea9b19e7613ba78db4552ecf9b204c056b4966d3fd691eb80b32" Dec 01 17:44:41 crc kubenswrapper[4868]: E1201 17:44:41.777304 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dbf91a124b9ea9b19e7613ba78db4552ecf9b204c056b4966d3fd691eb80b32\": container with ID starting with 5dbf91a124b9ea9b19e7613ba78db4552ecf9b204c056b4966d3fd691eb80b32 not found: ID does not exist" containerID="5dbf91a124b9ea9b19e7613ba78db4552ecf9b204c056b4966d3fd691eb80b32" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.777344 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dbf91a124b9ea9b19e7613ba78db4552ecf9b204c056b4966d3fd691eb80b32"} err="failed to get container status \"5dbf91a124b9ea9b19e7613ba78db4552ecf9b204c056b4966d3fd691eb80b32\": rpc error: code = NotFound desc = could not find container \"5dbf91a124b9ea9b19e7613ba78db4552ecf9b204c056b4966d3fd691eb80b32\": container with ID starting with 5dbf91a124b9ea9b19e7613ba78db4552ecf9b204c056b4966d3fd691eb80b32 not found: ID does not exist" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.777373 4868 scope.go:117] "RemoveContainer" containerID="98299ac37c9d398981045965e1ed7484fda45b7142252554239df2eab9ac0ba1" Dec 01 17:44:41 crc kubenswrapper[4868]: E1201 17:44:41.777926 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98299ac37c9d398981045965e1ed7484fda45b7142252554239df2eab9ac0ba1\": container with ID starting with 98299ac37c9d398981045965e1ed7484fda45b7142252554239df2eab9ac0ba1 not found: ID does not exist" containerID="98299ac37c9d398981045965e1ed7484fda45b7142252554239df2eab9ac0ba1" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.778012 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98299ac37c9d398981045965e1ed7484fda45b7142252554239df2eab9ac0ba1"} err="failed to get container status \"98299ac37c9d398981045965e1ed7484fda45b7142252554239df2eab9ac0ba1\": rpc error: code = NotFound desc = could not find container \"98299ac37c9d398981045965e1ed7484fda45b7142252554239df2eab9ac0ba1\": container with ID starting with 98299ac37c9d398981045965e1ed7484fda45b7142252554239df2eab9ac0ba1 not found: ID does not exist" Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.821371 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.840690 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-97868dbf6-djbqg"] Dec 01 17:44:41 crc kubenswrapper[4868]: I1201 17:44:41.846632 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-97868dbf6-djbqg"] Dec 01 17:44:42 crc kubenswrapper[4868]: I1201 17:44:42.183587 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdb719ec-3812-4252-bdb9-1bc93aa35462" path="/var/lib/kubelet/pods/fdb719ec-3812-4252-bdb9-1bc93aa35462/volumes" Dec 01 17:44:42 crc kubenswrapper[4868]: I1201 17:44:42.595279 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef3afc2e-b496-4104-98cf-35507acfe031","Type":"ContainerStarted","Data":"a94690676861fed3696b68306a268bc9ce8dff0c240e05b4b18e868e4da16742"} Dec 01 17:44:42 crc kubenswrapper[4868]: I1201 17:44:42.598401 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4247f042-96f0-4bc5-9ca3-a64ee5616624","Type":"ContainerStarted","Data":"f899d5e72bbfe1f48eaf9f5dc52a8091e5537587afc3ec691f262ad610d18553"} Dec 01 17:44:42 crc kubenswrapper[4868]: I1201 17:44:42.602091 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bc0d7b8-d497-4339-ad24-c9b75e6e06fc","Type":"ContainerStarted","Data":"2613ff883e4739097223ac002433f1c2eb83a605699f36a907604cd5686d4547"} Dec 01 17:44:42 crc kubenswrapper[4868]: I1201 17:44:42.638873 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.6388016610000005 podStartE2EDuration="4.638801661s" podCreationTimestamp="2025-12-01 17:44:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:44:42.636137298 +0000 UTC m=+1155.007247709" watchObservedRunningTime="2025-12-01 17:44:42.638801661 +0000 UTC m=+1155.009912072" Dec 01 17:44:44 crc kubenswrapper[4868]: I1201 17:44:44.629094 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef3afc2e-b496-4104-98cf-35507acfe031","Type":"ContainerStarted","Data":"a92b2364d5f0ec3b7ff48919b3046831e128d4e78c8890b88a3bbdc7891a98d0"} Dec 01 17:44:44 crc kubenswrapper[4868]: I1201 17:44:44.629649 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 17:44:44 crc kubenswrapper[4868]: I1201 17:44:44.629310 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="sg-core" containerID="cri-o://a94690676861fed3696b68306a268bc9ce8dff0c240e05b4b18e868e4da16742" gracePeriod=30 Dec 01 17:44:44 crc kubenswrapper[4868]: I1201 17:44:44.629351 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="ceilometer-notification-agent" containerID="cri-o://9be42b8f2e80c8df4a4ff4b89367f3467b0884b1e51b787b55055283d2ba8862" gracePeriod=30 Dec 01 17:44:44 crc kubenswrapper[4868]: I1201 17:44:44.629334 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="proxy-httpd" containerID="cri-o://a92b2364d5f0ec3b7ff48919b3046831e128d4e78c8890b88a3bbdc7891a98d0" gracePeriod=30 Dec 01 17:44:44 crc kubenswrapper[4868]: I1201 17:44:44.629286 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="ceilometer-central-agent" containerID="cri-o://3861fbd918cc8846b4d75999efd64c75f715a2f8ff83a10f1b6bff5fde69bb4e" gracePeriod=30 Dec 01 17:44:44 crc kubenswrapper[4868]: I1201 17:44:44.659645 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.596569382 podStartE2EDuration="7.659618963s" podCreationTimestamp="2025-12-01 17:44:37 +0000 UTC" firstStartedPulling="2025-12-01 17:44:39.307426552 +0000 UTC m=+1151.678536963" lastFinishedPulling="2025-12-01 17:44:43.370476133 +0000 UTC m=+1155.741586544" observedRunningTime="2025-12-01 17:44:44.656911659 +0000 UTC m=+1157.028022070" watchObservedRunningTime="2025-12-01 17:44:44.659618963 +0000 UTC m=+1157.030729374" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.391926 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k7gsb"] Dec 01 17:44:45 crc kubenswrapper[4868]: E1201 17:44:45.392737 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a996536c-2c87-4c80-90d6-c6758c109af1" containerName="mariadb-account-create-update" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.392760 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="a996536c-2c87-4c80-90d6-c6758c109af1" containerName="mariadb-account-create-update" Dec 01 17:44:45 crc kubenswrapper[4868]: E1201 17:44:45.392783 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="441fc0fe-f3fe-4e94-abd8-0d126b289042" containerName="mariadb-account-create-update" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.392791 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="441fc0fe-f3fe-4e94-abd8-0d126b289042" containerName="mariadb-account-create-update" Dec 01 17:44:45 crc kubenswrapper[4868]: E1201 17:44:45.392801 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb719ec-3812-4252-bdb9-1bc93aa35462" containerName="horizon-log" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.392807 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb719ec-3812-4252-bdb9-1bc93aa35462" containerName="horizon-log" Dec 01 17:44:45 crc kubenswrapper[4868]: E1201 17:44:45.392817 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb719ec-3812-4252-bdb9-1bc93aa35462" containerName="horizon" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.392823 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb719ec-3812-4252-bdb9-1bc93aa35462" containerName="horizon" Dec 01 17:44:45 crc kubenswrapper[4868]: E1201 17:44:45.392841 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b801bf8a-a39f-4537-aa83-06e958c31e02" containerName="mariadb-database-create" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.392846 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b801bf8a-a39f-4537-aa83-06e958c31e02" containerName="mariadb-database-create" Dec 01 17:44:45 crc kubenswrapper[4868]: E1201 17:44:45.392857 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5951a13-5101-4956-bcf6-345fab68ea75" containerName="mariadb-account-create-update" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.392862 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5951a13-5101-4956-bcf6-345fab68ea75" containerName="mariadb-account-create-update" Dec 01 17:44:45 crc kubenswrapper[4868]: E1201 17:44:45.392877 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d8af7d-f71e-42be-ab59-237f2d505b1f" containerName="mariadb-database-create" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.392883 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d8af7d-f71e-42be-ab59-237f2d505b1f" containerName="mariadb-database-create" Dec 01 17:44:45 crc kubenswrapper[4868]: E1201 17:44:45.392893 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="288e3235-6690-4c67-9e4f-2735be76d249" containerName="mariadb-database-create" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.392899 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="288e3235-6690-4c67-9e4f-2735be76d249" containerName="mariadb-database-create" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.393077 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="31d8af7d-f71e-42be-ab59-237f2d505b1f" containerName="mariadb-database-create" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.393090 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="441fc0fe-f3fe-4e94-abd8-0d126b289042" containerName="mariadb-account-create-update" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.393100 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="a996536c-2c87-4c80-90d6-c6758c109af1" containerName="mariadb-account-create-update" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.393109 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5951a13-5101-4956-bcf6-345fab68ea75" containerName="mariadb-account-create-update" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.393118 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="288e3235-6690-4c67-9e4f-2735be76d249" containerName="mariadb-database-create" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.393125 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdb719ec-3812-4252-bdb9-1bc93aa35462" containerName="horizon-log" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.393135 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdb719ec-3812-4252-bdb9-1bc93aa35462" containerName="horizon" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.393143 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="b801bf8a-a39f-4537-aa83-06e958c31e02" containerName="mariadb-database-create" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.393895 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.396245 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.396622 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.398313 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vcxcg" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.511138 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k7gsb"] Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.516280 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-scripts\") pod \"nova-cell0-conductor-db-sync-k7gsb\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.516437 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj6fh\" (UniqueName: \"kubernetes.io/projected/d0f4d962-9aa7-4ae9-836c-cdce14d53159-kube-api-access-mj6fh\") pod \"nova-cell0-conductor-db-sync-k7gsb\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.516559 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k7gsb\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.517522 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-config-data\") pod \"nova-cell0-conductor-db-sync-k7gsb\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.619138 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-config-data\") pod \"nova-cell0-conductor-db-sync-k7gsb\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.619200 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-scripts\") pod \"nova-cell0-conductor-db-sync-k7gsb\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.619236 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj6fh\" (UniqueName: \"kubernetes.io/projected/d0f4d962-9aa7-4ae9-836c-cdce14d53159-kube-api-access-mj6fh\") pod \"nova-cell0-conductor-db-sync-k7gsb\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.619274 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k7gsb\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.634878 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-config-data\") pod \"nova-cell0-conductor-db-sync-k7gsb\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.635285 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-scripts\") pod \"nova-cell0-conductor-db-sync-k7gsb\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.634977 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k7gsb\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.648042 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj6fh\" (UniqueName: \"kubernetes.io/projected/d0f4d962-9aa7-4ae9-836c-cdce14d53159-kube-api-access-mj6fh\") pod \"nova-cell0-conductor-db-sync-k7gsb\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.677735 4868 generic.go:334] "Generic (PLEG): container finished" podID="ef3afc2e-b496-4104-98cf-35507acfe031" containerID="a92b2364d5f0ec3b7ff48919b3046831e128d4e78c8890b88a3bbdc7891a98d0" exitCode=0 Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.677805 4868 generic.go:334] "Generic (PLEG): container finished" podID="ef3afc2e-b496-4104-98cf-35507acfe031" containerID="a94690676861fed3696b68306a268bc9ce8dff0c240e05b4b18e868e4da16742" exitCode=2 Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.677817 4868 generic.go:334] "Generic (PLEG): container finished" podID="ef3afc2e-b496-4104-98cf-35507acfe031" containerID="9be42b8f2e80c8df4a4ff4b89367f3467b0884b1e51b787b55055283d2ba8862" exitCode=0 Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.677878 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef3afc2e-b496-4104-98cf-35507acfe031","Type":"ContainerDied","Data":"a92b2364d5f0ec3b7ff48919b3046831e128d4e78c8890b88a3bbdc7891a98d0"} Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.677916 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef3afc2e-b496-4104-98cf-35507acfe031","Type":"ContainerDied","Data":"a94690676861fed3696b68306a268bc9ce8dff0c240e05b4b18e868e4da16742"} Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.677928 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef3afc2e-b496-4104-98cf-35507acfe031","Type":"ContainerDied","Data":"9be42b8f2e80c8df4a4ff4b89367f3467b0884b1e51b787b55055283d2ba8862"} Dec 01 17:44:45 crc kubenswrapper[4868]: I1201 17:44:45.799375 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:44:46 crc kubenswrapper[4868]: I1201 17:44:46.301164 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k7gsb"] Dec 01 17:44:46 crc kubenswrapper[4868]: W1201 17:44:46.305495 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0f4d962_9aa7_4ae9_836c_cdce14d53159.slice/crio-1cc8f5beaf1619c3db38be8128cf1ba032509822f24929fc3b3450ba3de8fdb0 WatchSource:0}: Error finding container 1cc8f5beaf1619c3db38be8128cf1ba032509822f24929fc3b3450ba3de8fdb0: Status 404 returned error can't find the container with id 1cc8f5beaf1619c3db38be8128cf1ba032509822f24929fc3b3450ba3de8fdb0 Dec 01 17:44:46 crc kubenswrapper[4868]: I1201 17:44:46.688275 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k7gsb" event={"ID":"d0f4d962-9aa7-4ae9-836c-cdce14d53159","Type":"ContainerStarted","Data":"1cc8f5beaf1619c3db38be8128cf1ba032509822f24929fc3b3450ba3de8fdb0"} Dec 01 17:44:47 crc kubenswrapper[4868]: I1201 17:44:47.707465 4868 generic.go:334] "Generic (PLEG): container finished" podID="ef3afc2e-b496-4104-98cf-35507acfe031" containerID="3861fbd918cc8846b4d75999efd64c75f715a2f8ff83a10f1b6bff5fde69bb4e" exitCode=0 Dec 01 17:44:47 crc kubenswrapper[4868]: I1201 17:44:47.707515 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef3afc2e-b496-4104-98cf-35507acfe031","Type":"ContainerDied","Data":"3861fbd918cc8846b4d75999efd64c75f715a2f8ff83a10f1b6bff5fde69bb4e"} Dec 01 17:44:47 crc kubenswrapper[4868]: I1201 17:44:47.987344 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.167381 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-combined-ca-bundle\") pod \"ef3afc2e-b496-4104-98cf-35507acfe031\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.167460 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef3afc2e-b496-4104-98cf-35507acfe031-run-httpd\") pod \"ef3afc2e-b496-4104-98cf-35507acfe031\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.168429 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef3afc2e-b496-4104-98cf-35507acfe031-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ef3afc2e-b496-4104-98cf-35507acfe031" (UID: "ef3afc2e-b496-4104-98cf-35507acfe031"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.168533 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-sg-core-conf-yaml\") pod \"ef3afc2e-b496-4104-98cf-35507acfe031\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.169053 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqhg4\" (UniqueName: \"kubernetes.io/projected/ef3afc2e-b496-4104-98cf-35507acfe031-kube-api-access-hqhg4\") pod \"ef3afc2e-b496-4104-98cf-35507acfe031\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.169126 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-scripts\") pod \"ef3afc2e-b496-4104-98cf-35507acfe031\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.169179 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-config-data\") pod \"ef3afc2e-b496-4104-98cf-35507acfe031\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.169225 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef3afc2e-b496-4104-98cf-35507acfe031-log-httpd\") pod \"ef3afc2e-b496-4104-98cf-35507acfe031\" (UID: \"ef3afc2e-b496-4104-98cf-35507acfe031\") " Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.170092 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef3afc2e-b496-4104-98cf-35507acfe031-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ef3afc2e-b496-4104-98cf-35507acfe031" (UID: "ef3afc2e-b496-4104-98cf-35507acfe031"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.170173 4868 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef3afc2e-b496-4104-98cf-35507acfe031-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.175237 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-scripts" (OuterVolumeSpecName: "scripts") pod "ef3afc2e-b496-4104-98cf-35507acfe031" (UID: "ef3afc2e-b496-4104-98cf-35507acfe031"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.207856 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef3afc2e-b496-4104-98cf-35507acfe031-kube-api-access-hqhg4" (OuterVolumeSpecName: "kube-api-access-hqhg4") pod "ef3afc2e-b496-4104-98cf-35507acfe031" (UID: "ef3afc2e-b496-4104-98cf-35507acfe031"). InnerVolumeSpecName "kube-api-access-hqhg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.216201 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ef3afc2e-b496-4104-98cf-35507acfe031" (UID: "ef3afc2e-b496-4104-98cf-35507acfe031"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.271453 4868 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.271489 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqhg4\" (UniqueName: \"kubernetes.io/projected/ef3afc2e-b496-4104-98cf-35507acfe031-kube-api-access-hqhg4\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.271501 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.271510 4868 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef3afc2e-b496-4104-98cf-35507acfe031-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.272761 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef3afc2e-b496-4104-98cf-35507acfe031" (UID: "ef3afc2e-b496-4104-98cf-35507acfe031"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.307469 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-config-data" (OuterVolumeSpecName: "config-data") pod "ef3afc2e-b496-4104-98cf-35507acfe031" (UID: "ef3afc2e-b496-4104-98cf-35507acfe031"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.376226 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.376527 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef3afc2e-b496-4104-98cf-35507acfe031-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.729861 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef3afc2e-b496-4104-98cf-35507acfe031","Type":"ContainerDied","Data":"f4627629b250f4a68ebef59261453e39d521b7684c91a1f883c2509678999b98"} Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.730627 4868 scope.go:117] "RemoveContainer" containerID="a92b2364d5f0ec3b7ff48919b3046831e128d4e78c8890b88a3bbdc7891a98d0" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.730110 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.771139 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.783436 4868 scope.go:117] "RemoveContainer" containerID="a94690676861fed3696b68306a268bc9ce8dff0c240e05b4b18e868e4da16742" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.792910 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.803075 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:44:48 crc kubenswrapper[4868]: E1201 17:44:48.825498 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="ceilometer-notification-agent" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.825548 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="ceilometer-notification-agent" Dec 01 17:44:48 crc kubenswrapper[4868]: E1201 17:44:48.825591 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="ceilometer-central-agent" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.825602 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="ceilometer-central-agent" Dec 01 17:44:48 crc kubenswrapper[4868]: E1201 17:44:48.825623 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="sg-core" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.825629 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="sg-core" Dec 01 17:44:48 crc kubenswrapper[4868]: E1201 17:44:48.825685 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="proxy-httpd" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.825693 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="proxy-httpd" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.826087 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="ceilometer-central-agent" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.826130 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="proxy-httpd" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.826142 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="ceilometer-notification-agent" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.827424 4868 scope.go:117] "RemoveContainer" containerID="9be42b8f2e80c8df4a4ff4b89367f3467b0884b1e51b787b55055283d2ba8862" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.841615 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" containerName="sg-core" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.844454 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.848396 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.849020 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.849582 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.874044 4868 scope.go:117] "RemoveContainer" containerID="3861fbd918cc8846b4d75999efd64c75f715a2f8ff83a10f1b6bff5fde69bb4e" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.985504 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.987297 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.991324 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-run-httpd\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.991423 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-config-data\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.991484 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-scripts\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.991505 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.991529 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wxpj\" (UniqueName: \"kubernetes.io/projected/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-kube-api-access-7wxpj\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.991546 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-log-httpd\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:48 crc kubenswrapper[4868]: I1201 17:44:48.991810 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.035778 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.069073 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.094278 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-run-httpd\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.094413 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-config-data\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.094483 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-scripts\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.094508 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.094531 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wxpj\" (UniqueName: \"kubernetes.io/projected/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-kube-api-access-7wxpj\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.094552 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-log-httpd\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.094646 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.094874 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-run-httpd\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.096188 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-log-httpd\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.104647 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.104929 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-scripts\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.106479 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-config-data\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.106715 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.124307 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wxpj\" (UniqueName: \"kubernetes.io/projected/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-kube-api-access-7wxpj\") pod \"ceilometer-0\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.200613 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.274772 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.274834 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.321118 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.363150 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.675389 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.743927 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f","Type":"ContainerStarted","Data":"a6b39db4a6752c1984823dcf12b97878af1ef9f8d7b32c4d7442ed604103667d"} Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.744025 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.744041 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.744369 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 17:44:49 crc kubenswrapper[4868]: I1201 17:44:49.744611 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 17:44:50 crc kubenswrapper[4868]: I1201 17:44:50.184028 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef3afc2e-b496-4104-98cf-35507acfe031" path="/var/lib/kubelet/pods/ef3afc2e-b496-4104-98cf-35507acfe031/volumes" Dec 01 17:44:50 crc kubenswrapper[4868]: I1201 17:44:50.693568 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:44:50 crc kubenswrapper[4868]: I1201 17:44:50.757043 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f","Type":"ContainerStarted","Data":"f3234b2fdd3db4f798c740cd16dc8e327f29ffcf4a2ff046aa31e3ba2854700c"} Dec 01 17:44:51 crc kubenswrapper[4868]: I1201 17:44:51.782500 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f","Type":"ContainerStarted","Data":"481340a260043a04f759e7664b0288fbdb088c2963f12d9d82d4e4217c4eefed"} Dec 01 17:44:51 crc kubenswrapper[4868]: I1201 17:44:51.847355 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 17:44:51 crc kubenswrapper[4868]: I1201 17:44:51.847527 4868 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 17:44:51 crc kubenswrapper[4868]: I1201 17:44:51.896608 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 17:44:51 crc kubenswrapper[4868]: I1201 17:44:51.968233 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 17:44:51 crc kubenswrapper[4868]: I1201 17:44:51.969109 4868 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 17:44:52 crc kubenswrapper[4868]: I1201 17:44:52.057502 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 17:44:58 crc kubenswrapper[4868]: I1201 17:44:58.872728 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k7gsb" event={"ID":"d0f4d962-9aa7-4ae9-836c-cdce14d53159","Type":"ContainerStarted","Data":"adeebb3b056bc9d1d5a5af8450952e30d08eff9e76732c29baa0606725dae518"} Dec 01 17:44:58 crc kubenswrapper[4868]: I1201 17:44:58.879732 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f","Type":"ContainerStarted","Data":"4e6ac636b2777f155406e336459f5b9f103b7008aa1903ba63cf1ebd72135a3c"} Dec 01 17:44:58 crc kubenswrapper[4868]: I1201 17:44:58.904618 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-k7gsb" podStartSLOduration=2.086658508 podStartE2EDuration="13.904549588s" podCreationTimestamp="2025-12-01 17:44:45 +0000 UTC" firstStartedPulling="2025-12-01 17:44:46.308755689 +0000 UTC m=+1158.679866100" lastFinishedPulling="2025-12-01 17:44:58.126646769 +0000 UTC m=+1170.497757180" observedRunningTime="2025-12-01 17:44:58.891453859 +0000 UTC m=+1171.262564280" watchObservedRunningTime="2025-12-01 17:44:58.904549588 +0000 UTC m=+1171.275659999" Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.149740 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls"] Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.151559 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.156995 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.158974 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.167499 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls"] Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.267956 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-config-volume\") pod \"collect-profiles-29410185-2m6ls\" (UID: \"7e91d120-4deb-4bc5-a4ae-b277d85ad10a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.268104 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvfkq\" (UniqueName: \"kubernetes.io/projected/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-kube-api-access-nvfkq\") pod \"collect-profiles-29410185-2m6ls\" (UID: \"7e91d120-4deb-4bc5-a4ae-b277d85ad10a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.268128 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-secret-volume\") pod \"collect-profiles-29410185-2m6ls\" (UID: \"7e91d120-4deb-4bc5-a4ae-b277d85ad10a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.371355 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvfkq\" (UniqueName: \"kubernetes.io/projected/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-kube-api-access-nvfkq\") pod \"collect-profiles-29410185-2m6ls\" (UID: \"7e91d120-4deb-4bc5-a4ae-b277d85ad10a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.371414 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-secret-volume\") pod \"collect-profiles-29410185-2m6ls\" (UID: \"7e91d120-4deb-4bc5-a4ae-b277d85ad10a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.371651 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-config-volume\") pod \"collect-profiles-29410185-2m6ls\" (UID: \"7e91d120-4deb-4bc5-a4ae-b277d85ad10a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.372910 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-config-volume\") pod \"collect-profiles-29410185-2m6ls\" (UID: \"7e91d120-4deb-4bc5-a4ae-b277d85ad10a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.385661 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-secret-volume\") pod \"collect-profiles-29410185-2m6ls\" (UID: \"7e91d120-4deb-4bc5-a4ae-b277d85ad10a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.390511 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvfkq\" (UniqueName: \"kubernetes.io/projected/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-kube-api-access-nvfkq\") pod \"collect-profiles-29410185-2m6ls\" (UID: \"7e91d120-4deb-4bc5-a4ae-b277d85ad10a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.515850 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.924321 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f","Type":"ContainerStarted","Data":"6f8613ed937964ebf12ba9b1736790ebe917dcfbc9b177d8ae12e62e96a0c3c6"} Dec 01 17:45:00 crc kubenswrapper[4868]: I1201 17:45:00.989299 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls"] Dec 01 17:45:00 crc kubenswrapper[4868]: W1201 17:45:00.992655 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e91d120_4deb_4bc5_a4ae_b277d85ad10a.slice/crio-a323eef8f64b08a4b75e1e368cfd1335ddbc9f71e6683a0173b5ed88b247d5c9 WatchSource:0}: Error finding container a323eef8f64b08a4b75e1e368cfd1335ddbc9f71e6683a0173b5ed88b247d5c9: Status 404 returned error can't find the container with id a323eef8f64b08a4b75e1e368cfd1335ddbc9f71e6683a0173b5ed88b247d5c9 Dec 01 17:45:01 crc kubenswrapper[4868]: I1201 17:45:01.934510 4868 generic.go:334] "Generic (PLEG): container finished" podID="7e91d120-4deb-4bc5-a4ae-b277d85ad10a" containerID="9e450d322b643e6d9bcb8713507f4ad95b98cbf83227044c5a06618af0adff69" exitCode=0 Dec 01 17:45:01 crc kubenswrapper[4868]: I1201 17:45:01.934668 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" event={"ID":"7e91d120-4deb-4bc5-a4ae-b277d85ad10a","Type":"ContainerDied","Data":"9e450d322b643e6d9bcb8713507f4ad95b98cbf83227044c5a06618af0adff69"} Dec 01 17:45:01 crc kubenswrapper[4868]: I1201 17:45:01.935274 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" event={"ID":"7e91d120-4deb-4bc5-a4ae-b277d85ad10a","Type":"ContainerStarted","Data":"a323eef8f64b08a4b75e1e368cfd1335ddbc9f71e6683a0173b5ed88b247d5c9"} Dec 01 17:45:01 crc kubenswrapper[4868]: I1201 17:45:01.935452 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="ceilometer-central-agent" containerID="cri-o://f3234b2fdd3db4f798c740cd16dc8e327f29ffcf4a2ff046aa31e3ba2854700c" gracePeriod=30 Dec 01 17:45:01 crc kubenswrapper[4868]: I1201 17:45:01.935553 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="ceilometer-notification-agent" containerID="cri-o://481340a260043a04f759e7664b0288fbdb088c2963f12d9d82d4e4217c4eefed" gracePeriod=30 Dec 01 17:45:01 crc kubenswrapper[4868]: I1201 17:45:01.935558 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 17:45:01 crc kubenswrapper[4868]: I1201 17:45:01.935590 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="proxy-httpd" containerID="cri-o://6f8613ed937964ebf12ba9b1736790ebe917dcfbc9b177d8ae12e62e96a0c3c6" gracePeriod=30 Dec 01 17:45:01 crc kubenswrapper[4868]: I1201 17:45:01.935751 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="sg-core" containerID="cri-o://4e6ac636b2777f155406e336459f5b9f103b7008aa1903ba63cf1ebd72135a3c" gracePeriod=30 Dec 01 17:45:01 crc kubenswrapper[4868]: I1201 17:45:01.999450 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.1010230659999998 podStartE2EDuration="13.999429447s" podCreationTimestamp="2025-12-01 17:44:48 +0000 UTC" firstStartedPulling="2025-12-01 17:44:49.693399808 +0000 UTC m=+1162.064510219" lastFinishedPulling="2025-12-01 17:45:00.591806189 +0000 UTC m=+1172.962916600" observedRunningTime="2025-12-01 17:45:01.994407929 +0000 UTC m=+1174.365518340" watchObservedRunningTime="2025-12-01 17:45:01.999429447 +0000 UTC m=+1174.370539858" Dec 01 17:45:02 crc kubenswrapper[4868]: E1201 17:45:02.139345 4868 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6dca2f2_be0b_4c58_a8a5_bf5810b72a9f.slice/crio-4e6ac636b2777f155406e336459f5b9f103b7008aa1903ba63cf1ebd72135a3c.scope\": RecentStats: unable to find data in memory cache]" Dec 01 17:45:02 crc kubenswrapper[4868]: I1201 17:45:02.999263 4868 generic.go:334] "Generic (PLEG): container finished" podID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerID="6f8613ed937964ebf12ba9b1736790ebe917dcfbc9b177d8ae12e62e96a0c3c6" exitCode=0 Dec 01 17:45:02 crc kubenswrapper[4868]: I1201 17:45:02.999836 4868 generic.go:334] "Generic (PLEG): container finished" podID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerID="4e6ac636b2777f155406e336459f5b9f103b7008aa1903ba63cf1ebd72135a3c" exitCode=2 Dec 01 17:45:02 crc kubenswrapper[4868]: I1201 17:45:02.999846 4868 generic.go:334] "Generic (PLEG): container finished" podID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerID="481340a260043a04f759e7664b0288fbdb088c2963f12d9d82d4e4217c4eefed" exitCode=0 Dec 01 17:45:02 crc kubenswrapper[4868]: I1201 17:45:02.999853 4868 generic.go:334] "Generic (PLEG): container finished" podID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerID="f3234b2fdd3db4f798c740cd16dc8e327f29ffcf4a2ff046aa31e3ba2854700c" exitCode=0 Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.000150 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f","Type":"ContainerDied","Data":"6f8613ed937964ebf12ba9b1736790ebe917dcfbc9b177d8ae12e62e96a0c3c6"} Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.000204 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f","Type":"ContainerDied","Data":"4e6ac636b2777f155406e336459f5b9f103b7008aa1903ba63cf1ebd72135a3c"} Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.000217 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f","Type":"ContainerDied","Data":"481340a260043a04f759e7664b0288fbdb088c2963f12d9d82d4e4217c4eefed"} Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.000226 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f","Type":"ContainerDied","Data":"f3234b2fdd3db4f798c740cd16dc8e327f29ffcf4a2ff046aa31e3ba2854700c"} Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.157422 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.325138 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.329765 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-config-data\") pod \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.329851 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wxpj\" (UniqueName: \"kubernetes.io/projected/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-kube-api-access-7wxpj\") pod \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.329908 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-scripts\") pod \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.330036 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-run-httpd\") pod \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.330188 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-sg-core-conf-yaml\") pod \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.330218 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-combined-ca-bundle\") pod \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.330317 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-log-httpd\") pod \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\" (UID: \"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f\") " Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.331882 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" (UID: "f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.332111 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" (UID: "f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.339002 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-scripts" (OuterVolumeSpecName: "scripts") pod "f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" (UID: "f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.343339 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-kube-api-access-7wxpj" (OuterVolumeSpecName: "kube-api-access-7wxpj") pod "f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" (UID: "f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f"). InnerVolumeSpecName "kube-api-access-7wxpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.366285 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" (UID: "f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.426486 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-config-data" (OuterVolumeSpecName: "config-data") pod "f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" (UID: "f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.432306 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-secret-volume\") pod \"7e91d120-4deb-4bc5-a4ae-b277d85ad10a\" (UID: \"7e91d120-4deb-4bc5-a4ae-b277d85ad10a\") " Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.432515 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-config-volume\") pod \"7e91d120-4deb-4bc5-a4ae-b277d85ad10a\" (UID: \"7e91d120-4deb-4bc5-a4ae-b277d85ad10a\") " Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.432569 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvfkq\" (UniqueName: \"kubernetes.io/projected/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-kube-api-access-nvfkq\") pod \"7e91d120-4deb-4bc5-a4ae-b277d85ad10a\" (UID: \"7e91d120-4deb-4bc5-a4ae-b277d85ad10a\") " Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.434004 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-config-volume" (OuterVolumeSpecName: "config-volume") pod "7e91d120-4deb-4bc5-a4ae-b277d85ad10a" (UID: "7e91d120-4deb-4bc5-a4ae-b277d85ad10a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.435979 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.436012 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wxpj\" (UniqueName: \"kubernetes.io/projected/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-kube-api-access-7wxpj\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.436027 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.436039 4868 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.436051 4868 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.436061 4868 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.436071 4868 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.436135 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-kube-api-access-nvfkq" (OuterVolumeSpecName: "kube-api-access-nvfkq") pod "7e91d120-4deb-4bc5-a4ae-b277d85ad10a" (UID: "7e91d120-4deb-4bc5-a4ae-b277d85ad10a"). InnerVolumeSpecName "kube-api-access-nvfkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.443621 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7e91d120-4deb-4bc5-a4ae-b277d85ad10a" (UID: "7e91d120-4deb-4bc5-a4ae-b277d85ad10a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.455384 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" (UID: "f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.538187 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.538231 4868 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:03 crc kubenswrapper[4868]: I1201 17:45:03.538241 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvfkq\" (UniqueName: \"kubernetes.io/projected/7e91d120-4deb-4bc5-a4ae-b277d85ad10a-kube-api-access-nvfkq\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.011674 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" event={"ID":"7e91d120-4deb-4bc5-a4ae-b277d85ad10a","Type":"ContainerDied","Data":"a323eef8f64b08a4b75e1e368cfd1335ddbc9f71e6683a0173b5ed88b247d5c9"} Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.011728 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a323eef8f64b08a4b75e1e368cfd1335ddbc9f71e6683a0173b5ed88b247d5c9" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.011805 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.020824 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f","Type":"ContainerDied","Data":"a6b39db4a6752c1984823dcf12b97878af1ef9f8d7b32c4d7442ed604103667d"} Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.021073 4868 scope.go:117] "RemoveContainer" containerID="6f8613ed937964ebf12ba9b1736790ebe917dcfbc9b177d8ae12e62e96a0c3c6" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.020874 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.047532 4868 scope.go:117] "RemoveContainer" containerID="4e6ac636b2777f155406e336459f5b9f103b7008aa1903ba63cf1ebd72135a3c" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.064107 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.084813 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.106185 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:45:04 crc kubenswrapper[4868]: E1201 17:45:04.106715 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="ceilometer-notification-agent" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.106731 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="ceilometer-notification-agent" Dec 01 17:45:04 crc kubenswrapper[4868]: E1201 17:45:04.106745 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="sg-core" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.106751 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="sg-core" Dec 01 17:45:04 crc kubenswrapper[4868]: E1201 17:45:04.106772 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="ceilometer-central-agent" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.106783 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="ceilometer-central-agent" Dec 01 17:45:04 crc kubenswrapper[4868]: E1201 17:45:04.106803 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e91d120-4deb-4bc5-a4ae-b277d85ad10a" containerName="collect-profiles" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.106809 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e91d120-4deb-4bc5-a4ae-b277d85ad10a" containerName="collect-profiles" Dec 01 17:45:04 crc kubenswrapper[4868]: E1201 17:45:04.106832 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="proxy-httpd" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.106837 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="proxy-httpd" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.107070 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e91d120-4deb-4bc5-a4ae-b277d85ad10a" containerName="collect-profiles" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.107096 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="ceilometer-central-agent" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.107109 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="proxy-httpd" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.107119 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="ceilometer-notification-agent" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.107131 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" containerName="sg-core" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.109134 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.114364 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.116558 4868 scope.go:117] "RemoveContainer" containerID="481340a260043a04f759e7664b0288fbdb088c2963f12d9d82d4e4217c4eefed" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.117415 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.133907 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.150613 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-scripts\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.150699 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.150718 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-run-httpd\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.150828 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-log-httpd\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.151060 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.151106 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-config-data\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.151154 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdqvf\" (UniqueName: \"kubernetes.io/projected/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-kube-api-access-jdqvf\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.168045 4868 scope.go:117] "RemoveContainer" containerID="f3234b2fdd3db4f798c740cd16dc8e327f29ffcf4a2ff046aa31e3ba2854700c" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.183473 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f" path="/var/lib/kubelet/pods/f6dca2f2-be0b-4c58-a8a5-bf5810b72a9f/volumes" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.252905 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-scripts\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.253011 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.253028 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-run-httpd\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.253046 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-log-httpd\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.253099 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.253120 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-config-data\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.253139 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdqvf\" (UniqueName: \"kubernetes.io/projected/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-kube-api-access-jdqvf\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.253910 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-run-httpd\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.254411 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-log-httpd\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.257176 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.257374 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.257629 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-config-data\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.259481 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-scripts\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.272558 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdqvf\" (UniqueName: \"kubernetes.io/projected/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-kube-api-access-jdqvf\") pod \"ceilometer-0\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.439465 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:45:04 crc kubenswrapper[4868]: I1201 17:45:04.926859 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:45:05 crc kubenswrapper[4868]: I1201 17:45:05.030421 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cda1e79-77f9-4b20-a0cf-cf8f0619228b","Type":"ContainerStarted","Data":"1ff4d68691ffcdf9c4f6f58546fabc2c5a133b4a74a3eab4af21225187113530"} Dec 01 17:45:06 crc kubenswrapper[4868]: I1201 17:45:06.040783 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cda1e79-77f9-4b20-a0cf-cf8f0619228b","Type":"ContainerStarted","Data":"26ba2d03acdd9505144fa1012002e04da97c9687cdffc890667f92030f2421b0"} Dec 01 17:45:07 crc kubenswrapper[4868]: I1201 17:45:07.060542 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cda1e79-77f9-4b20-a0cf-cf8f0619228b","Type":"ContainerStarted","Data":"dfdeae37b27c6cdc9c4b5428bb5dd3a1a91f78c21054d7315effb2577bf49b91"} Dec 01 17:45:08 crc kubenswrapper[4868]: I1201 17:45:08.072635 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cda1e79-77f9-4b20-a0cf-cf8f0619228b","Type":"ContainerStarted","Data":"32aae52f5b745b711e8c82229ce99775a3b15bb0c9dbb65806e5a29ed6efad33"} Dec 01 17:45:10 crc kubenswrapper[4868]: I1201 17:45:10.091258 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k7gsb" event={"ID":"d0f4d962-9aa7-4ae9-836c-cdce14d53159","Type":"ContainerDied","Data":"adeebb3b056bc9d1d5a5af8450952e30d08eff9e76732c29baa0606725dae518"} Dec 01 17:45:10 crc kubenswrapper[4868]: I1201 17:45:10.091190 4868 generic.go:334] "Generic (PLEG): container finished" podID="d0f4d962-9aa7-4ae9-836c-cdce14d53159" containerID="adeebb3b056bc9d1d5a5af8450952e30d08eff9e76732c29baa0606725dae518" exitCode=0 Dec 01 17:45:10 crc kubenswrapper[4868]: I1201 17:45:10.106576 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cda1e79-77f9-4b20-a0cf-cf8f0619228b","Type":"ContainerStarted","Data":"90efad9f1a86acadc5d0089dd3923d5a9bff46b9f8cd7e5b7bc64bf9d023349f"} Dec 01 17:45:10 crc kubenswrapper[4868]: I1201 17:45:10.150632 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.668992153 podStartE2EDuration="6.150612877s" podCreationTimestamp="2025-12-01 17:45:04 +0000 UTC" firstStartedPulling="2025-12-01 17:45:04.945268059 +0000 UTC m=+1177.316378470" lastFinishedPulling="2025-12-01 17:45:09.426888783 +0000 UTC m=+1181.797999194" observedRunningTime="2025-12-01 17:45:10.150163875 +0000 UTC m=+1182.521274286" watchObservedRunningTime="2025-12-01 17:45:10.150612877 +0000 UTC m=+1182.521723288" Dec 01 17:45:11 crc kubenswrapper[4868]: I1201 17:45:11.117684 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 17:45:11 crc kubenswrapper[4868]: I1201 17:45:11.471930 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:45:11 crc kubenswrapper[4868]: I1201 17:45:11.540464 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-scripts\") pod \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " Dec 01 17:45:11 crc kubenswrapper[4868]: I1201 17:45:11.540608 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj6fh\" (UniqueName: \"kubernetes.io/projected/d0f4d962-9aa7-4ae9-836c-cdce14d53159-kube-api-access-mj6fh\") pod \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " Dec 01 17:45:11 crc kubenswrapper[4868]: I1201 17:45:11.540731 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-combined-ca-bundle\") pod \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " Dec 01 17:45:11 crc kubenswrapper[4868]: I1201 17:45:11.540852 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-config-data\") pod \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\" (UID: \"d0f4d962-9aa7-4ae9-836c-cdce14d53159\") " Dec 01 17:45:11 crc kubenswrapper[4868]: I1201 17:45:11.546593 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-scripts" (OuterVolumeSpecName: "scripts") pod "d0f4d962-9aa7-4ae9-836c-cdce14d53159" (UID: "d0f4d962-9aa7-4ae9-836c-cdce14d53159"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:11 crc kubenswrapper[4868]: I1201 17:45:11.546750 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0f4d962-9aa7-4ae9-836c-cdce14d53159-kube-api-access-mj6fh" (OuterVolumeSpecName: "kube-api-access-mj6fh") pod "d0f4d962-9aa7-4ae9-836c-cdce14d53159" (UID: "d0f4d962-9aa7-4ae9-836c-cdce14d53159"). InnerVolumeSpecName "kube-api-access-mj6fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:45:11 crc kubenswrapper[4868]: I1201 17:45:11.566769 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-config-data" (OuterVolumeSpecName: "config-data") pod "d0f4d962-9aa7-4ae9-836c-cdce14d53159" (UID: "d0f4d962-9aa7-4ae9-836c-cdce14d53159"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:11 crc kubenswrapper[4868]: I1201 17:45:11.575658 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0f4d962-9aa7-4ae9-836c-cdce14d53159" (UID: "d0f4d962-9aa7-4ae9-836c-cdce14d53159"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:11 crc kubenswrapper[4868]: I1201 17:45:11.643037 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:11 crc kubenswrapper[4868]: I1201 17:45:11.643078 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:11 crc kubenswrapper[4868]: I1201 17:45:11.643087 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f4d962-9aa7-4ae9-836c-cdce14d53159-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:11 crc kubenswrapper[4868]: I1201 17:45:11.643097 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj6fh\" (UniqueName: \"kubernetes.io/projected/d0f4d962-9aa7-4ae9-836c-cdce14d53159-kube-api-access-mj6fh\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.131061 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k7gsb" event={"ID":"d0f4d962-9aa7-4ae9-836c-cdce14d53159","Type":"ContainerDied","Data":"1cc8f5beaf1619c3db38be8128cf1ba032509822f24929fc3b3450ba3de8fdb0"} Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.131122 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cc8f5beaf1619c3db38be8128cf1ba032509822f24929fc3b3450ba3de8fdb0" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.131076 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k7gsb" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.233325 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 17:45:12 crc kubenswrapper[4868]: E1201 17:45:12.233823 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0f4d962-9aa7-4ae9-836c-cdce14d53159" containerName="nova-cell0-conductor-db-sync" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.233843 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0f4d962-9aa7-4ae9-836c-cdce14d53159" containerName="nova-cell0-conductor-db-sync" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.234043 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0f4d962-9aa7-4ae9-836c-cdce14d53159" containerName="nova-cell0-conductor-db-sync" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.234720 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.237531 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vcxcg" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.237566 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.250097 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.253057 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a0dfb3-863c-4c80-b46c-e4e3b5fe4098-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"83a0dfb3-863c-4c80-b46c-e4e3b5fe4098\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.253124 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a0dfb3-863c-4c80-b46c-e4e3b5fe4098-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"83a0dfb3-863c-4c80-b46c-e4e3b5fe4098\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.253261 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf9gs\" (UniqueName: \"kubernetes.io/projected/83a0dfb3-863c-4c80-b46c-e4e3b5fe4098-kube-api-access-rf9gs\") pod \"nova-cell0-conductor-0\" (UID: \"83a0dfb3-863c-4c80-b46c-e4e3b5fe4098\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.354581 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a0dfb3-863c-4c80-b46c-e4e3b5fe4098-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"83a0dfb3-863c-4c80-b46c-e4e3b5fe4098\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.354633 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a0dfb3-863c-4c80-b46c-e4e3b5fe4098-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"83a0dfb3-863c-4c80-b46c-e4e3b5fe4098\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.354714 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf9gs\" (UniqueName: \"kubernetes.io/projected/83a0dfb3-863c-4c80-b46c-e4e3b5fe4098-kube-api-access-rf9gs\") pod \"nova-cell0-conductor-0\" (UID: \"83a0dfb3-863c-4c80-b46c-e4e3b5fe4098\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.358760 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a0dfb3-863c-4c80-b46c-e4e3b5fe4098-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"83a0dfb3-863c-4c80-b46c-e4e3b5fe4098\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.369272 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a0dfb3-863c-4c80-b46c-e4e3b5fe4098-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"83a0dfb3-863c-4c80-b46c-e4e3b5fe4098\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.369696 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf9gs\" (UniqueName: \"kubernetes.io/projected/83a0dfb3-863c-4c80-b46c-e4e3b5fe4098-kube-api-access-rf9gs\") pod \"nova-cell0-conductor-0\" (UID: \"83a0dfb3-863c-4c80-b46c-e4e3b5fe4098\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:45:12 crc kubenswrapper[4868]: I1201 17:45:12.573898 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 17:45:13 crc kubenswrapper[4868]: I1201 17:45:13.014433 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 17:45:13 crc kubenswrapper[4868]: W1201 17:45:13.018094 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a0dfb3_863c_4c80_b46c_e4e3b5fe4098.slice/crio-261375aea1d3ebcacce71a4cf472749d492eb8dda9d93f3f53902bc6ea44989d WatchSource:0}: Error finding container 261375aea1d3ebcacce71a4cf472749d492eb8dda9d93f3f53902bc6ea44989d: Status 404 returned error can't find the container with id 261375aea1d3ebcacce71a4cf472749d492eb8dda9d93f3f53902bc6ea44989d Dec 01 17:45:13 crc kubenswrapper[4868]: I1201 17:45:13.140511 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"83a0dfb3-863c-4c80-b46c-e4e3b5fe4098","Type":"ContainerStarted","Data":"261375aea1d3ebcacce71a4cf472749d492eb8dda9d93f3f53902bc6ea44989d"} Dec 01 17:45:14 crc kubenswrapper[4868]: I1201 17:45:14.151546 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"83a0dfb3-863c-4c80-b46c-e4e3b5fe4098","Type":"ContainerStarted","Data":"d02fe1c080dd87fb3f842c109a327236079993fdd6d2f0ac7a923582b0cc6857"} Dec 01 17:45:14 crc kubenswrapper[4868]: I1201 17:45:14.153428 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 01 17:45:22 crc kubenswrapper[4868]: I1201 17:45:22.606011 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 01 17:45:22 crc kubenswrapper[4868]: I1201 17:45:22.631822 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=10.631772464 podStartE2EDuration="10.631772464s" podCreationTimestamp="2025-12-01 17:45:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:45:14.174361372 +0000 UTC m=+1186.545471783" watchObservedRunningTime="2025-12-01 17:45:22.631772464 +0000 UTC m=+1195.002882885" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.060468 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-2tbrq"] Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.061873 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.065837 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.066217 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.075734 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-2tbrq"] Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.175512 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2tbrq\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.175679 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-config-data\") pod \"nova-cell0-cell-mapping-2tbrq\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.175743 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxpgq\" (UniqueName: \"kubernetes.io/projected/fb439a97-8909-414a-a4cf-22e94e909356-kube-api-access-rxpgq\") pod \"nova-cell0-cell-mapping-2tbrq\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.175845 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-scripts\") pod \"nova-cell0-cell-mapping-2tbrq\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.265928 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.267478 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.270554 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.276972 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxpgq\" (UniqueName: \"kubernetes.io/projected/fb439a97-8909-414a-a4cf-22e94e909356-kube-api-access-rxpgq\") pod \"nova-cell0-cell-mapping-2tbrq\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.277054 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-scripts\") pod \"nova-cell0-cell-mapping-2tbrq\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.277162 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2tbrq\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.277243 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-config-data\") pod \"nova-cell0-cell-mapping-2tbrq\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.283884 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-scripts\") pod \"nova-cell0-cell-mapping-2tbrq\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.285536 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2tbrq\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.285727 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-config-data\") pod \"nova-cell0-cell-mapping-2tbrq\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.290385 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.322805 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxpgq\" (UniqueName: \"kubernetes.io/projected/fb439a97-8909-414a-a4cf-22e94e909356-kube-api-access-rxpgq\") pod \"nova-cell0-cell-mapping-2tbrq\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.370825 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.372306 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.375184 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.378882 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6v57\" (UniqueName: \"kubernetes.io/projected/ba81079d-403c-45fa-b29d-df13d2d927d8-kube-api-access-k6v57\") pod \"nova-scheduler-0\" (UID: \"ba81079d-403c-45fa-b29d-df13d2d927d8\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.378974 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba81079d-403c-45fa-b29d-df13d2d927d8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ba81079d-403c-45fa-b29d-df13d2d927d8\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.379004 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba81079d-403c-45fa-b29d-df13d2d927d8-config-data\") pod \"nova-scheduler-0\" (UID: \"ba81079d-403c-45fa-b29d-df13d2d927d8\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.383541 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.386310 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.436799 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.438437 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.453237 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.455750 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.482076 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1bc0fc1-0afa-4e05-abe0-05285c85576f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " pod="openstack/nova-api-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.482132 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba81079d-403c-45fa-b29d-df13d2d927d8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ba81079d-403c-45fa-b29d-df13d2d927d8\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.482248 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1bc0fc1-0afa-4e05-abe0-05285c85576f-config-data\") pod \"nova-api-0\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " pod="openstack/nova-api-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.482305 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba81079d-403c-45fa-b29d-df13d2d927d8-config-data\") pod \"nova-scheduler-0\" (UID: \"ba81079d-403c-45fa-b29d-df13d2d927d8\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.482430 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgdgv\" (UniqueName: \"kubernetes.io/projected/a1bc0fc1-0afa-4e05-abe0-05285c85576f-kube-api-access-pgdgv\") pod \"nova-api-0\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " pod="openstack/nova-api-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.482509 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1bc0fc1-0afa-4e05-abe0-05285c85576f-logs\") pod \"nova-api-0\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " pod="openstack/nova-api-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.482651 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6v57\" (UniqueName: \"kubernetes.io/projected/ba81079d-403c-45fa-b29d-df13d2d927d8-kube-api-access-k6v57\") pod \"nova-scheduler-0\" (UID: \"ba81079d-403c-45fa-b29d-df13d2d927d8\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.485338 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba81079d-403c-45fa-b29d-df13d2d927d8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ba81079d-403c-45fa-b29d-df13d2d927d8\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.498621 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba81079d-403c-45fa-b29d-df13d2d927d8-config-data\") pod \"nova-scheduler-0\" (UID: \"ba81079d-403c-45fa-b29d-df13d2d927d8\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.520528 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6v57\" (UniqueName: \"kubernetes.io/projected/ba81079d-403c-45fa-b29d-df13d2d927d8-kube-api-access-k6v57\") pod \"nova-scheduler-0\" (UID: \"ba81079d-403c-45fa-b29d-df13d2d927d8\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.563009 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.564630 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.573954 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.585320 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1bc0fc1-0afa-4e05-abe0-05285c85576f-config-data\") pod \"nova-api-0\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " pod="openstack/nova-api-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.585411 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgdgv\" (UniqueName: \"kubernetes.io/projected/a1bc0fc1-0afa-4e05-abe0-05285c85576f-kube-api-access-pgdgv\") pod \"nova-api-0\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " pod="openstack/nova-api-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.585446 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1bc0fc1-0afa-4e05-abe0-05285c85576f-logs\") pod \"nova-api-0\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " pod="openstack/nova-api-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.585477 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-logs\") pod \"nova-metadata-0\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " pod="openstack/nova-metadata-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.585505 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-config-data\") pod \"nova-metadata-0\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " pod="openstack/nova-metadata-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.585572 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1bc0fc1-0afa-4e05-abe0-05285c85576f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " pod="openstack/nova-api-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.585588 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " pod="openstack/nova-metadata-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.585615 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfdjb\" (UniqueName: \"kubernetes.io/projected/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-kube-api-access-zfdjb\") pod \"nova-metadata-0\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " pod="openstack/nova-metadata-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.587156 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1bc0fc1-0afa-4e05-abe0-05285c85576f-logs\") pod \"nova-api-0\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " pod="openstack/nova-api-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.587460 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.591580 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1bc0fc1-0afa-4e05-abe0-05285c85576f-config-data\") pod \"nova-api-0\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " pod="openstack/nova-api-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.595606 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1bc0fc1-0afa-4e05-abe0-05285c85576f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " pod="openstack/nova-api-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.618911 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-z2bdk"] Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.631342 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.650617 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgdgv\" (UniqueName: \"kubernetes.io/projected/a1bc0fc1-0afa-4e05-abe0-05285c85576f-kube-api-access-pgdgv\") pod \"nova-api-0\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " pod="openstack/nova-api-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.668211 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-z2bdk"] Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.687624 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-logs\") pod \"nova-metadata-0\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " pod="openstack/nova-metadata-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.687670 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-config-data\") pod \"nova-metadata-0\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " pod="openstack/nova-metadata-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.687742 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e3e9efc3-aa56-442a-9e27-44b72f11d3ad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.687777 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " pod="openstack/nova-metadata-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.687801 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfdjb\" (UniqueName: \"kubernetes.io/projected/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-kube-api-access-zfdjb\") pod \"nova-metadata-0\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " pod="openstack/nova-metadata-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.687820 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfxz4\" (UniqueName: \"kubernetes.io/projected/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-kube-api-access-xfxz4\") pod \"nova-cell1-novncproxy-0\" (UID: \"e3e9efc3-aa56-442a-9e27-44b72f11d3ad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.687842 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e3e9efc3-aa56-442a-9e27-44b72f11d3ad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.692372 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-logs\") pod \"nova-metadata-0\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " pod="openstack/nova-metadata-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.699230 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-config-data\") pod \"nova-metadata-0\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " pod="openstack/nova-metadata-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.699474 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " pod="openstack/nova-metadata-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.710388 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfdjb\" (UniqueName: \"kubernetes.io/projected/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-kube-api-access-zfdjb\") pod \"nova-metadata-0\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " pod="openstack/nova-metadata-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.731800 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.789305 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-dns-svc\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.789398 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfxz4\" (UniqueName: \"kubernetes.io/projected/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-kube-api-access-xfxz4\") pod \"nova-cell1-novncproxy-0\" (UID: \"e3e9efc3-aa56-442a-9e27-44b72f11d3ad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.789429 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e3e9efc3-aa56-442a-9e27-44b72f11d3ad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.789465 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.789536 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.789597 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c9dl\" (UniqueName: \"kubernetes.io/projected/19e9c57b-8750-414f-b95b-3938c1356ed0-kube-api-access-5c9dl\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.789654 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-config\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.789721 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.789761 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e3e9efc3-aa56-442a-9e27-44b72f11d3ad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.796110 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e3e9efc3-aa56-442a-9e27-44b72f11d3ad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.797491 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e3e9efc3-aa56-442a-9e27-44b72f11d3ad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.815701 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfxz4\" (UniqueName: \"kubernetes.io/projected/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-kube-api-access-xfxz4\") pod \"nova-cell1-novncproxy-0\" (UID: \"e3e9efc3-aa56-442a-9e27-44b72f11d3ad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.892480 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.892886 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-dns-svc\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.892981 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.893141 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.893215 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c9dl\" (UniqueName: \"kubernetes.io/projected/19e9c57b-8750-414f-b95b-3938c1356ed0-kube-api-access-5c9dl\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.893256 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-config\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.893468 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.894155 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.894707 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-dns-svc\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.894739 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-config\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.895126 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.911005 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c9dl\" (UniqueName: \"kubernetes.io/projected/19e9c57b-8750-414f-b95b-3938c1356ed0-kube-api-access-5c9dl\") pod \"dnsmasq-dns-bccf8f775-z2bdk\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.919610 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.977855 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:45:23 crc kubenswrapper[4868]: I1201 17:45:23.994816 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.021180 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.086144 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-2tbrq"] Dec 01 17:45:24 crc kubenswrapper[4868]: W1201 17:45:24.101549 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb439a97_8909_414a_a4cf_22e94e909356.slice/crio-908cba80c131d8c933f9513de8123958a41a544fa8ecdc2aa13df85cb4ff9674 WatchSource:0}: Error finding container 908cba80c131d8c933f9513de8123958a41a544fa8ecdc2aa13df85cb4ff9674: Status 404 returned error can't find the container with id 908cba80c131d8c933f9513de8123958a41a544fa8ecdc2aa13df85cb4ff9674 Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.257307 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2tbrq" event={"ID":"fb439a97-8909-414a-a4cf-22e94e909356","Type":"ContainerStarted","Data":"908cba80c131d8c933f9513de8123958a41a544fa8ecdc2aa13df85cb4ff9674"} Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.306915 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.346663 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5pwdf"] Dec 01 17:45:24 crc kubenswrapper[4868]: W1201 17:45:24.348197 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba81079d_403c_45fa_b29d_df13d2d927d8.slice/crio-066ff1affb5f83b9e7743ae4e41783ff4178240facb8ad9307ba3309fd76c145 WatchSource:0}: Error finding container 066ff1affb5f83b9e7743ae4e41783ff4178240facb8ad9307ba3309fd76c145: Status 404 returned error can't find the container with id 066ff1affb5f83b9e7743ae4e41783ff4178240facb8ad9307ba3309fd76c145 Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.355502 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.360834 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.363108 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.399194 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5pwdf"] Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.431335 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.521461 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plpsh\" (UniqueName: \"kubernetes.io/projected/897da32f-2dbb-416b-b58e-a4d59040c89c-kube-api-access-plpsh\") pod \"nova-cell1-conductor-db-sync-5pwdf\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.521880 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5pwdf\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.521950 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-scripts\") pod \"nova-cell1-conductor-db-sync-5pwdf\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.522039 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-config-data\") pod \"nova-cell1-conductor-db-sync-5pwdf\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.623600 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plpsh\" (UniqueName: \"kubernetes.io/projected/897da32f-2dbb-416b-b58e-a4d59040c89c-kube-api-access-plpsh\") pod \"nova-cell1-conductor-db-sync-5pwdf\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.623677 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5pwdf\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.623721 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-scripts\") pod \"nova-cell1-conductor-db-sync-5pwdf\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.623810 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-config-data\") pod \"nova-cell1-conductor-db-sync-5pwdf\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.634783 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-config-data\") pod \"nova-cell1-conductor-db-sync-5pwdf\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.644180 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.646642 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5pwdf\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.647213 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-scripts\") pod \"nova-cell1-conductor-db-sync-5pwdf\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.650139 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plpsh\" (UniqueName: \"kubernetes.io/projected/897da32f-2dbb-416b-b58e-a4d59040c89c-kube-api-access-plpsh\") pod \"nova-cell1-conductor-db-sync-5pwdf\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:24 crc kubenswrapper[4868]: W1201 17:45:24.670588 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc9ee16e_96a0_4bda_9b2a_2c1b50019b58.slice/crio-41835642667b368c81d1d7e31abe72f511067f8b46ce5a80007b799048921295 WatchSource:0}: Error finding container 41835642667b368c81d1d7e31abe72f511067f8b46ce5a80007b799048921295: Status 404 returned error can't find the container with id 41835642667b368c81d1d7e31abe72f511067f8b46ce5a80007b799048921295 Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.764142 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.778517 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-z2bdk"] Dec 01 17:45:24 crc kubenswrapper[4868]: I1201 17:45:24.788838 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:25 crc kubenswrapper[4868]: I1201 17:45:25.255470 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5pwdf"] Dec 01 17:45:25 crc kubenswrapper[4868]: W1201 17:45:25.267144 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod897da32f_2dbb_416b_b58e_a4d59040c89c.slice/crio-1cbcd7b0b67a11074b0418314ed5e6f99ac17ebcb4865ac5f8a21cbf4bf4fddd WatchSource:0}: Error finding container 1cbcd7b0b67a11074b0418314ed5e6f99ac17ebcb4865ac5f8a21cbf4bf4fddd: Status 404 returned error can't find the container with id 1cbcd7b0b67a11074b0418314ed5e6f99ac17ebcb4865ac5f8a21cbf4bf4fddd Dec 01 17:45:25 crc kubenswrapper[4868]: I1201 17:45:25.268328 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e3e9efc3-aa56-442a-9e27-44b72f11d3ad","Type":"ContainerStarted","Data":"bdbf7468d1357f01e99e6fc2c7a79da4a4a4eb28dccd45c9a40607321b71a44e"} Dec 01 17:45:25 crc kubenswrapper[4868]: I1201 17:45:25.270700 4868 generic.go:334] "Generic (PLEG): container finished" podID="19e9c57b-8750-414f-b95b-3938c1356ed0" containerID="f15b7bfacdd6842acbe87c872c21527f51a023616f5ae43cfb4af45011ae642e" exitCode=0 Dec 01 17:45:25 crc kubenswrapper[4868]: I1201 17:45:25.270804 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" event={"ID":"19e9c57b-8750-414f-b95b-3938c1356ed0","Type":"ContainerDied","Data":"f15b7bfacdd6842acbe87c872c21527f51a023616f5ae43cfb4af45011ae642e"} Dec 01 17:45:25 crc kubenswrapper[4868]: I1201 17:45:25.270871 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" event={"ID":"19e9c57b-8750-414f-b95b-3938c1356ed0","Type":"ContainerStarted","Data":"d410126333bdad254e9d4c334cef6d36bbac6eed618d2dc6a58eac27f2145a11"} Dec 01 17:45:25 crc kubenswrapper[4868]: I1201 17:45:25.275109 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ba81079d-403c-45fa-b29d-df13d2d927d8","Type":"ContainerStarted","Data":"066ff1affb5f83b9e7743ae4e41783ff4178240facb8ad9307ba3309fd76c145"} Dec 01 17:45:25 crc kubenswrapper[4868]: I1201 17:45:25.276986 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a1bc0fc1-0afa-4e05-abe0-05285c85576f","Type":"ContainerStarted","Data":"2246a3aafd627aa81555347369cc896943681daa6266504618279404b5a000e8"} Dec 01 17:45:25 crc kubenswrapper[4868]: I1201 17:45:25.280263 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58","Type":"ContainerStarted","Data":"41835642667b368c81d1d7e31abe72f511067f8b46ce5a80007b799048921295"} Dec 01 17:45:25 crc kubenswrapper[4868]: I1201 17:45:25.283153 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2tbrq" event={"ID":"fb439a97-8909-414a-a4cf-22e94e909356","Type":"ContainerStarted","Data":"84f8dc40aac9df502e5e1ba67b84436cd7f524979e2974092d7a86954bb89e34"} Dec 01 17:45:25 crc kubenswrapper[4868]: I1201 17:45:25.333363 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-2tbrq" podStartSLOduration=2.333344012 podStartE2EDuration="2.333344012s" podCreationTimestamp="2025-12-01 17:45:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:45:25.323732039 +0000 UTC m=+1197.694842450" watchObservedRunningTime="2025-12-01 17:45:25.333344012 +0000 UTC m=+1197.704454423" Dec 01 17:45:25 crc kubenswrapper[4868]: I1201 17:45:25.905391 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:45:25 crc kubenswrapper[4868]: I1201 17:45:25.905705 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:45:26 crc kubenswrapper[4868]: I1201 17:45:26.303507 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5pwdf" event={"ID":"897da32f-2dbb-416b-b58e-a4d59040c89c","Type":"ContainerStarted","Data":"e2237ad0b6d8c1771316a95bd844ac047784f7c1e7107f6a7b93de8775347385"} Dec 01 17:45:26 crc kubenswrapper[4868]: I1201 17:45:26.303555 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5pwdf" event={"ID":"897da32f-2dbb-416b-b58e-a4d59040c89c","Type":"ContainerStarted","Data":"1cbcd7b0b67a11074b0418314ed5e6f99ac17ebcb4865ac5f8a21cbf4bf4fddd"} Dec 01 17:45:26 crc kubenswrapper[4868]: I1201 17:45:26.313100 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" event={"ID":"19e9c57b-8750-414f-b95b-3938c1356ed0","Type":"ContainerStarted","Data":"68c50850b354708778d1c5f0bf21257e977a75918b3a2edeb8bcb5d0ae9e3f23"} Dec 01 17:45:26 crc kubenswrapper[4868]: I1201 17:45:26.313366 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:26 crc kubenswrapper[4868]: I1201 17:45:26.335915 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-5pwdf" podStartSLOduration=2.335880598 podStartE2EDuration="2.335880598s" podCreationTimestamp="2025-12-01 17:45:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:45:26.320903858 +0000 UTC m=+1198.692014269" watchObservedRunningTime="2025-12-01 17:45:26.335880598 +0000 UTC m=+1198.706991009" Dec 01 17:45:26 crc kubenswrapper[4868]: I1201 17:45:26.345783 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" podStartSLOduration=3.3457617490000002 podStartE2EDuration="3.345761749s" podCreationTimestamp="2025-12-01 17:45:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:45:26.344304249 +0000 UTC m=+1198.715414660" watchObservedRunningTime="2025-12-01 17:45:26.345761749 +0000 UTC m=+1198.716872160" Dec 01 17:45:27 crc kubenswrapper[4868]: I1201 17:45:27.123521 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 17:45:27 crc kubenswrapper[4868]: I1201 17:45:27.137856 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:45:29 crc kubenswrapper[4868]: I1201 17:45:29.365303 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a1bc0fc1-0afa-4e05-abe0-05285c85576f","Type":"ContainerStarted","Data":"e84d1798bc0ae4641d48befea34912af24a042cf4624eb73c2fe18000183365e"} Dec 01 17:45:29 crc kubenswrapper[4868]: I1201 17:45:29.366372 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a1bc0fc1-0afa-4e05-abe0-05285c85576f","Type":"ContainerStarted","Data":"581755dab82fc21797870f98cdbd213bc13b37d56d056a1c91852285b1b4e4db"} Dec 01 17:45:29 crc kubenswrapper[4868]: I1201 17:45:29.367381 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ba81079d-403c-45fa-b29d-df13d2d927d8","Type":"ContainerStarted","Data":"2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f"} Dec 01 17:45:29 crc kubenswrapper[4868]: I1201 17:45:29.369783 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58","Type":"ContainerStarted","Data":"3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490"} Dec 01 17:45:29 crc kubenswrapper[4868]: I1201 17:45:29.369861 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fc9ee16e-96a0-4bda-9b2a-2c1b50019b58" containerName="nova-metadata-log" containerID="cri-o://db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e" gracePeriod=30 Dec 01 17:45:29 crc kubenswrapper[4868]: I1201 17:45:29.369925 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fc9ee16e-96a0-4bda-9b2a-2c1b50019b58" containerName="nova-metadata-metadata" containerID="cri-o://3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490" gracePeriod=30 Dec 01 17:45:29 crc kubenswrapper[4868]: I1201 17:45:29.369874 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58","Type":"ContainerStarted","Data":"db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e"} Dec 01 17:45:29 crc kubenswrapper[4868]: I1201 17:45:29.371838 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e3e9efc3-aa56-442a-9e27-44b72f11d3ad","Type":"ContainerStarted","Data":"8aad16845f92ec5322db4c77da6ca3cceb72bc33cbede6390eace86eb1d9829f"} Dec 01 17:45:29 crc kubenswrapper[4868]: I1201 17:45:29.372016 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="e3e9efc3-aa56-442a-9e27-44b72f11d3ad" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://8aad16845f92ec5322db4c77da6ca3cceb72bc33cbede6390eace86eb1d9829f" gracePeriod=30 Dec 01 17:45:29 crc kubenswrapper[4868]: I1201 17:45:29.391330 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.607444194 podStartE2EDuration="6.391308975s" podCreationTimestamp="2025-12-01 17:45:23 +0000 UTC" firstStartedPulling="2025-12-01 17:45:24.441883611 +0000 UTC m=+1196.812994022" lastFinishedPulling="2025-12-01 17:45:28.225748392 +0000 UTC m=+1200.596858803" observedRunningTime="2025-12-01 17:45:29.390169093 +0000 UTC m=+1201.761279504" watchObservedRunningTime="2025-12-01 17:45:29.391308975 +0000 UTC m=+1201.762419386" Dec 01 17:45:29 crc kubenswrapper[4868]: I1201 17:45:29.423280 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.558710919 podStartE2EDuration="6.42325648s" podCreationTimestamp="2025-12-01 17:45:23 +0000 UTC" firstStartedPulling="2025-12-01 17:45:24.360654885 +0000 UTC m=+1196.731765296" lastFinishedPulling="2025-12-01 17:45:28.225200446 +0000 UTC m=+1200.596310857" observedRunningTime="2025-12-01 17:45:29.413249715 +0000 UTC m=+1201.784360126" watchObservedRunningTime="2025-12-01 17:45:29.42325648 +0000 UTC m=+1201.794366901" Dec 01 17:45:29 crc kubenswrapper[4868]: I1201 17:45:29.433202 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.968699786 podStartE2EDuration="6.433180602s" podCreationTimestamp="2025-12-01 17:45:23 +0000 UTC" firstStartedPulling="2025-12-01 17:45:24.760439602 +0000 UTC m=+1197.131550013" lastFinishedPulling="2025-12-01 17:45:28.224920418 +0000 UTC m=+1200.596030829" observedRunningTime="2025-12-01 17:45:29.432905305 +0000 UTC m=+1201.804015726" watchObservedRunningTime="2025-12-01 17:45:29.433180602 +0000 UTC m=+1201.804291023" Dec 01 17:45:29 crc kubenswrapper[4868]: I1201 17:45:29.459686 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.896879157 podStartE2EDuration="6.459521274s" podCreationTimestamp="2025-12-01 17:45:23 +0000 UTC" firstStartedPulling="2025-12-01 17:45:24.675485743 +0000 UTC m=+1197.046596154" lastFinishedPulling="2025-12-01 17:45:28.23812786 +0000 UTC m=+1200.609238271" observedRunningTime="2025-12-01 17:45:29.451158405 +0000 UTC m=+1201.822268826" watchObservedRunningTime="2025-12-01 17:45:29.459521274 +0000 UTC m=+1201.830631695" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.021883 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.177829 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-config-data\") pod \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.177918 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfdjb\" (UniqueName: \"kubernetes.io/projected/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-kube-api-access-zfdjb\") pod \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.177980 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-logs\") pod \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.178025 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-combined-ca-bundle\") pod \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\" (UID: \"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58\") " Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.178682 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-logs" (OuterVolumeSpecName: "logs") pod "fc9ee16e-96a0-4bda-9b2a-2c1b50019b58" (UID: "fc9ee16e-96a0-4bda-9b2a-2c1b50019b58"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.185620 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-kube-api-access-zfdjb" (OuterVolumeSpecName: "kube-api-access-zfdjb") pod "fc9ee16e-96a0-4bda-9b2a-2c1b50019b58" (UID: "fc9ee16e-96a0-4bda-9b2a-2c1b50019b58"). InnerVolumeSpecName "kube-api-access-zfdjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.211220 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-config-data" (OuterVolumeSpecName: "config-data") pod "fc9ee16e-96a0-4bda-9b2a-2c1b50019b58" (UID: "fc9ee16e-96a0-4bda-9b2a-2c1b50019b58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.214220 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc9ee16e-96a0-4bda-9b2a-2c1b50019b58" (UID: "fc9ee16e-96a0-4bda-9b2a-2c1b50019b58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.282095 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.282134 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfdjb\" (UniqueName: \"kubernetes.io/projected/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-kube-api-access-zfdjb\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.282145 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.282157 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.381302 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.387109 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58","Type":"ContainerDied","Data":"3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490"} Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.387204 4868 scope.go:117] "RemoveContainer" containerID="3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.387048 4868 generic.go:334] "Generic (PLEG): container finished" podID="fc9ee16e-96a0-4bda-9b2a-2c1b50019b58" containerID="3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490" exitCode=0 Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.388669 4868 generic.go:334] "Generic (PLEG): container finished" podID="fc9ee16e-96a0-4bda-9b2a-2c1b50019b58" containerID="db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e" exitCode=143 Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.388868 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58","Type":"ContainerDied","Data":"db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e"} Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.388934 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc9ee16e-96a0-4bda-9b2a-2c1b50019b58","Type":"ContainerDied","Data":"41835642667b368c81d1d7e31abe72f511067f8b46ce5a80007b799048921295"} Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.428174 4868 scope.go:117] "RemoveContainer" containerID="db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.443189 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.460001 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.462325 4868 scope.go:117] "RemoveContainer" containerID="3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490" Dec 01 17:45:30 crc kubenswrapper[4868]: E1201 17:45:30.464381 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490\": container with ID starting with 3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490 not found: ID does not exist" containerID="3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.464445 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490"} err="failed to get container status \"3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490\": rpc error: code = NotFound desc = could not find container \"3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490\": container with ID starting with 3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490 not found: ID does not exist" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.464479 4868 scope.go:117] "RemoveContainer" containerID="db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e" Dec 01 17:45:30 crc kubenswrapper[4868]: E1201 17:45:30.465098 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e\": container with ID starting with db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e not found: ID does not exist" containerID="db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.465174 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e"} err="failed to get container status \"db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e\": rpc error: code = NotFound desc = could not find container \"db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e\": container with ID starting with db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e not found: ID does not exist" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.465210 4868 scope.go:117] "RemoveContainer" containerID="3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.465683 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490"} err="failed to get container status \"3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490\": rpc error: code = NotFound desc = could not find container \"3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490\": container with ID starting with 3828397ed32e1dd732b1ca25dbb49a3ba7d0b2fc2e7ec1ab1d4c6be8bb3be490 not found: ID does not exist" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.465729 4868 scope.go:117] "RemoveContainer" containerID="db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.466182 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e"} err="failed to get container status \"db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e\": rpc error: code = NotFound desc = could not find container \"db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e\": container with ID starting with db455a93cffa6f210cb7fac8e88385b162dc36c3226180ef8e9a01868685252e not found: ID does not exist" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.472109 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:45:30 crc kubenswrapper[4868]: E1201 17:45:30.472592 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc9ee16e-96a0-4bda-9b2a-2c1b50019b58" containerName="nova-metadata-metadata" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.472610 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc9ee16e-96a0-4bda-9b2a-2c1b50019b58" containerName="nova-metadata-metadata" Dec 01 17:45:30 crc kubenswrapper[4868]: E1201 17:45:30.472640 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc9ee16e-96a0-4bda-9b2a-2c1b50019b58" containerName="nova-metadata-log" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.472647 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc9ee16e-96a0-4bda-9b2a-2c1b50019b58" containerName="nova-metadata-log" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.472871 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc9ee16e-96a0-4bda-9b2a-2c1b50019b58" containerName="nova-metadata-log" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.472894 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc9ee16e-96a0-4bda-9b2a-2c1b50019b58" containerName="nova-metadata-metadata" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.474332 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.477461 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.477727 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.485494 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.593196 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbtw4\" (UniqueName: \"kubernetes.io/projected/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-kube-api-access-nbtw4\") pod \"nova-metadata-0\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.593265 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.593285 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-logs\") pod \"nova-metadata-0\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.593305 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.593890 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-config-data\") pod \"nova-metadata-0\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.695617 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-config-data\") pod \"nova-metadata-0\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.695731 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbtw4\" (UniqueName: \"kubernetes.io/projected/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-kube-api-access-nbtw4\") pod \"nova-metadata-0\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.695775 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.695795 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-logs\") pod \"nova-metadata-0\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.695816 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.697090 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-logs\") pod \"nova-metadata-0\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.702670 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.703407 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.711582 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbtw4\" (UniqueName: \"kubernetes.io/projected/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-kube-api-access-nbtw4\") pod \"nova-metadata-0\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.721579 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-config-data\") pod \"nova-metadata-0\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " pod="openstack/nova-metadata-0" Dec 01 17:45:30 crc kubenswrapper[4868]: I1201 17:45:30.800181 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:45:31 crc kubenswrapper[4868]: I1201 17:45:31.309415 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:45:31 crc kubenswrapper[4868]: I1201 17:45:31.402735 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7","Type":"ContainerStarted","Data":"50e2aaebb30c9f170d935b992825d17a3519ea687c7f73a88ed42edc75fe2d4e"} Dec 01 17:45:32 crc kubenswrapper[4868]: I1201 17:45:32.186973 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc9ee16e-96a0-4bda-9b2a-2c1b50019b58" path="/var/lib/kubelet/pods/fc9ee16e-96a0-4bda-9b2a-2c1b50019b58/volumes" Dec 01 17:45:32 crc kubenswrapper[4868]: I1201 17:45:32.423398 4868 generic.go:334] "Generic (PLEG): container finished" podID="fb439a97-8909-414a-a4cf-22e94e909356" containerID="84f8dc40aac9df502e5e1ba67b84436cd7f524979e2974092d7a86954bb89e34" exitCode=0 Dec 01 17:45:32 crc kubenswrapper[4868]: I1201 17:45:32.423463 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2tbrq" event={"ID":"fb439a97-8909-414a-a4cf-22e94e909356","Type":"ContainerDied","Data":"84f8dc40aac9df502e5e1ba67b84436cd7f524979e2974092d7a86954bb89e34"} Dec 01 17:45:32 crc kubenswrapper[4868]: I1201 17:45:32.431730 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7","Type":"ContainerStarted","Data":"3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3"} Dec 01 17:45:32 crc kubenswrapper[4868]: I1201 17:45:32.432141 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7","Type":"ContainerStarted","Data":"eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b"} Dec 01 17:45:32 crc kubenswrapper[4868]: I1201 17:45:32.476933 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.476907828 podStartE2EDuration="2.476907828s" podCreationTimestamp="2025-12-01 17:45:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:45:32.473626538 +0000 UTC m=+1204.844736959" watchObservedRunningTime="2025-12-01 17:45:32.476907828 +0000 UTC m=+1204.848018239" Dec 01 17:45:32 crc kubenswrapper[4868]: E1201 17:45:32.917379 4868 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod897da32f_2dbb_416b_b58e_a4d59040c89c.slice/crio-conmon-e2237ad0b6d8c1771316a95bd844ac047784f7c1e7107f6a7b93de8775347385.scope\": RecentStats: unable to find data in memory cache]" Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.460136 4868 generic.go:334] "Generic (PLEG): container finished" podID="897da32f-2dbb-416b-b58e-a4d59040c89c" containerID="e2237ad0b6d8c1771316a95bd844ac047784f7c1e7107f6a7b93de8775347385" exitCode=0 Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.461455 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5pwdf" event={"ID":"897da32f-2dbb-416b-b58e-a4d59040c89c","Type":"ContainerDied","Data":"e2237ad0b6d8c1771316a95bd844ac047784f7c1e7107f6a7b93de8775347385"} Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.732740 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.733097 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.773197 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.819469 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.922168 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.922464 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.963803 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-combined-ca-bundle\") pod \"fb439a97-8909-414a-a4cf-22e94e909356\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.964317 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxpgq\" (UniqueName: \"kubernetes.io/projected/fb439a97-8909-414a-a4cf-22e94e909356-kube-api-access-rxpgq\") pod \"fb439a97-8909-414a-a4cf-22e94e909356\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.964499 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-config-data\") pod \"fb439a97-8909-414a-a4cf-22e94e909356\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.964634 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-scripts\") pod \"fb439a97-8909-414a-a4cf-22e94e909356\" (UID: \"fb439a97-8909-414a-a4cf-22e94e909356\") " Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.973093 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-scripts" (OuterVolumeSpecName: "scripts") pod "fb439a97-8909-414a-a4cf-22e94e909356" (UID: "fb439a97-8909-414a-a4cf-22e94e909356"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.973191 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb439a97-8909-414a-a4cf-22e94e909356-kube-api-access-rxpgq" (OuterVolumeSpecName: "kube-api-access-rxpgq") pod "fb439a97-8909-414a-a4cf-22e94e909356" (UID: "fb439a97-8909-414a-a4cf-22e94e909356"). InnerVolumeSpecName "kube-api-access-rxpgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.992214 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb439a97-8909-414a-a4cf-22e94e909356" (UID: "fb439a97-8909-414a-a4cf-22e94e909356"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.995319 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:45:33 crc kubenswrapper[4868]: I1201 17:45:33.999847 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-config-data" (OuterVolumeSpecName: "config-data") pod "fb439a97-8909-414a-a4cf-22e94e909356" (UID: "fb439a97-8909-414a-a4cf-22e94e909356"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.023895 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.072394 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.072431 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.072440 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb439a97-8909-414a-a4cf-22e94e909356-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.072451 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxpgq\" (UniqueName: \"kubernetes.io/projected/fb439a97-8909-414a-a4cf-22e94e909356-kube-api-access-rxpgq\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.098824 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-cfjj4"] Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.100653 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" podUID="69c7a38b-c21b-46cd-a120-1d4cd1867941" containerName="dnsmasq-dns" containerID="cri-o://331aeb57154733165d0f6615693a045df1bd6043a1081c6b748d58c0c2bbd31a" gracePeriod=10 Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.453150 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.472540 4868 generic.go:334] "Generic (PLEG): container finished" podID="69c7a38b-c21b-46cd-a120-1d4cd1867941" containerID="331aeb57154733165d0f6615693a045df1bd6043a1081c6b748d58c0c2bbd31a" exitCode=0 Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.472613 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" event={"ID":"69c7a38b-c21b-46cd-a120-1d4cd1867941","Type":"ContainerDied","Data":"331aeb57154733165d0f6615693a045df1bd6043a1081c6b748d58c0c2bbd31a"} Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.476073 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2tbrq" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.476606 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2tbrq" event={"ID":"fb439a97-8909-414a-a4cf-22e94e909356","Type":"ContainerDied","Data":"908cba80c131d8c933f9513de8123958a41a544fa8ecdc2aa13df85cb4ff9674"} Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.476628 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="908cba80c131d8c933f9513de8123958a41a544fa8ecdc2aa13df85cb4ff9674" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.518565 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.554745 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.688519 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sj6sb\" (UniqueName: \"kubernetes.io/projected/69c7a38b-c21b-46cd-a120-1d4cd1867941-kube-api-access-sj6sb\") pod \"69c7a38b-c21b-46cd-a120-1d4cd1867941\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.688605 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-ovsdbserver-nb\") pod \"69c7a38b-c21b-46cd-a120-1d4cd1867941\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.688679 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-ovsdbserver-sb\") pod \"69c7a38b-c21b-46cd-a120-1d4cd1867941\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.688771 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-dns-svc\") pod \"69c7a38b-c21b-46cd-a120-1d4cd1867941\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.688824 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-config\") pod \"69c7a38b-c21b-46cd-a120-1d4cd1867941\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.688859 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-dns-swift-storage-0\") pod \"69c7a38b-c21b-46cd-a120-1d4cd1867941\" (UID: \"69c7a38b-c21b-46cd-a120-1d4cd1867941\") " Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.696789 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69c7a38b-c21b-46cd-a120-1d4cd1867941-kube-api-access-sj6sb" (OuterVolumeSpecName: "kube-api-access-sj6sb") pod "69c7a38b-c21b-46cd-a120-1d4cd1867941" (UID: "69c7a38b-c21b-46cd-a120-1d4cd1867941"). InnerVolumeSpecName "kube-api-access-sj6sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.752380 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "69c7a38b-c21b-46cd-a120-1d4cd1867941" (UID: "69c7a38b-c21b-46cd-a120-1d4cd1867941"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.764683 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.764873 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a1bc0fc1-0afa-4e05-abe0-05285c85576f" containerName="nova-api-log" containerID="cri-o://581755dab82fc21797870f98cdbd213bc13b37d56d056a1c91852285b1b4e4db" gracePeriod=30 Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.765512 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a1bc0fc1-0afa-4e05-abe0-05285c85576f" containerName="nova-api-api" containerID="cri-o://e84d1798bc0ae4641d48befea34912af24a042cf4624eb73c2fe18000183365e" gracePeriod=30 Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.777707 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a1bc0fc1-0afa-4e05-abe0-05285c85576f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": EOF" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.777707 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a1bc0fc1-0afa-4e05-abe0-05285c85576f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": EOF" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.792231 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.792410 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sj6sb\" (UniqueName: \"kubernetes.io/projected/69c7a38b-c21b-46cd-a120-1d4cd1867941-kube-api-access-sj6sb\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.849573 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "69c7a38b-c21b-46cd-a120-1d4cd1867941" (UID: "69c7a38b-c21b-46cd-a120-1d4cd1867941"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.851557 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "69c7a38b-c21b-46cd-a120-1d4cd1867941" (UID: "69c7a38b-c21b-46cd-a120-1d4cd1867941"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.859500 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "69c7a38b-c21b-46cd-a120-1d4cd1867941" (UID: "69c7a38b-c21b-46cd-a120-1d4cd1867941"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.864110 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.864308 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" containerName="nova-metadata-log" containerID="cri-o://eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b" gracePeriod=30 Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.864727 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" containerName="nova-metadata-metadata" containerID="cri-o://3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3" gracePeriod=30 Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.880723 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-config" (OuterVolumeSpecName: "config") pod "69c7a38b-c21b-46cd-a120-1d4cd1867941" (UID: "69c7a38b-c21b-46cd-a120-1d4cd1867941"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.897379 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.897413 4868 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.897428 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.897440 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69c7a38b-c21b-46cd-a120-1d4cd1867941-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:34 crc kubenswrapper[4868]: I1201 17:45:34.948264 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.067963 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.100365 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-combined-ca-bundle\") pod \"897da32f-2dbb-416b-b58e-a4d59040c89c\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.100525 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-config-data\") pod \"897da32f-2dbb-416b-b58e-a4d59040c89c\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.100638 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plpsh\" (UniqueName: \"kubernetes.io/projected/897da32f-2dbb-416b-b58e-a4d59040c89c-kube-api-access-plpsh\") pod \"897da32f-2dbb-416b-b58e-a4d59040c89c\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.100702 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-scripts\") pod \"897da32f-2dbb-416b-b58e-a4d59040c89c\" (UID: \"897da32f-2dbb-416b-b58e-a4d59040c89c\") " Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.104062 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/897da32f-2dbb-416b-b58e-a4d59040c89c-kube-api-access-plpsh" (OuterVolumeSpecName: "kube-api-access-plpsh") pod "897da32f-2dbb-416b-b58e-a4d59040c89c" (UID: "897da32f-2dbb-416b-b58e-a4d59040c89c"). InnerVolumeSpecName "kube-api-access-plpsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.105584 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-scripts" (OuterVolumeSpecName: "scripts") pod "897da32f-2dbb-416b-b58e-a4d59040c89c" (UID: "897da32f-2dbb-416b-b58e-a4d59040c89c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.126293 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "897da32f-2dbb-416b-b58e-a4d59040c89c" (UID: "897da32f-2dbb-416b-b58e-a4d59040c89c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.136108 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-config-data" (OuterVolumeSpecName: "config-data") pod "897da32f-2dbb-416b-b58e-a4d59040c89c" (UID: "897da32f-2dbb-416b-b58e-a4d59040c89c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.202614 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plpsh\" (UniqueName: \"kubernetes.io/projected/897da32f-2dbb-416b-b58e-a4d59040c89c-kube-api-access-plpsh\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.202650 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.202664 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.202676 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/897da32f-2dbb-416b-b58e-a4d59040c89c-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.325391 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.405718 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-logs\") pod \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.405798 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-config-data\") pod \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.406027 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbtw4\" (UniqueName: \"kubernetes.io/projected/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-kube-api-access-nbtw4\") pod \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.406234 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-nova-metadata-tls-certs\") pod \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.406295 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-combined-ca-bundle\") pod \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\" (UID: \"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7\") " Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.406595 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-logs" (OuterVolumeSpecName: "logs") pod "9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" (UID: "9fb1678e-e191-43d1-a9a9-d50b83dfe8f7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.406861 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.413418 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-kube-api-access-nbtw4" (OuterVolumeSpecName: "kube-api-access-nbtw4") pod "9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" (UID: "9fb1678e-e191-43d1-a9a9-d50b83dfe8f7"). InnerVolumeSpecName "kube-api-access-nbtw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.438787 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" (UID: "9fb1678e-e191-43d1-a9a9-d50b83dfe8f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.454167 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-config-data" (OuterVolumeSpecName: "config-data") pod "9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" (UID: "9fb1678e-e191-43d1-a9a9-d50b83dfe8f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.465331 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" (UID: "9fb1678e-e191-43d1-a9a9-d50b83dfe8f7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.486158 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" event={"ID":"69c7a38b-c21b-46cd-a120-1d4cd1867941","Type":"ContainerDied","Data":"f52cc5e86f2c22fc91faa0fbb6c3af111cf5af50496bbd2f78be42de9cf55a9f"} Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.486222 4868 scope.go:117] "RemoveContainer" containerID="331aeb57154733165d0f6615693a045df1bd6043a1081c6b748d58c0c2bbd31a" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.486365 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-cfjj4" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.497824 4868 generic.go:334] "Generic (PLEG): container finished" podID="a1bc0fc1-0afa-4e05-abe0-05285c85576f" containerID="581755dab82fc21797870f98cdbd213bc13b37d56d056a1c91852285b1b4e4db" exitCode=143 Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.497927 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a1bc0fc1-0afa-4e05-abe0-05285c85576f","Type":"ContainerDied","Data":"581755dab82fc21797870f98cdbd213bc13b37d56d056a1c91852285b1b4e4db"} Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.506739 4868 generic.go:334] "Generic (PLEG): container finished" podID="9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" containerID="3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3" exitCode=0 Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.506782 4868 generic.go:334] "Generic (PLEG): container finished" podID="9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" containerID="eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b" exitCode=143 Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.506826 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.506847 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7","Type":"ContainerDied","Data":"3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3"} Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.506895 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7","Type":"ContainerDied","Data":"eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b"} Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.506910 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9fb1678e-e191-43d1-a9a9-d50b83dfe8f7","Type":"ContainerDied","Data":"50e2aaebb30c9f170d935b992825d17a3519ea687c7f73a88ed42edc75fe2d4e"} Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.508201 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbtw4\" (UniqueName: \"kubernetes.io/projected/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-kube-api-access-nbtw4\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.508239 4868 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.508253 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.508264 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.515908 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5pwdf" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.516709 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5pwdf" event={"ID":"897da32f-2dbb-416b-b58e-a4d59040c89c","Type":"ContainerDied","Data":"1cbcd7b0b67a11074b0418314ed5e6f99ac17ebcb4865ac5f8a21cbf4bf4fddd"} Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.516788 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cbcd7b0b67a11074b0418314ed5e6f99ac17ebcb4865ac5f8a21cbf4bf4fddd" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.523436 4868 scope.go:117] "RemoveContainer" containerID="6de604e3d047646f1da44a4148d0834d2d8d4dbaf250356cb700961f490323ec" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.571651 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-cfjj4"] Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.586019 4868 scope.go:117] "RemoveContainer" containerID="3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.608367 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-cfjj4"] Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.627405 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.631285 4868 scope.go:117] "RemoveContainer" containerID="eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.644368 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.655110 4868 scope.go:117] "RemoveContainer" containerID="3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3" Dec 01 17:45:35 crc kubenswrapper[4868]: E1201 17:45:35.660025 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3\": container with ID starting with 3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3 not found: ID does not exist" containerID="3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.660060 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3"} err="failed to get container status \"3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3\": rpc error: code = NotFound desc = could not find container \"3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3\": container with ID starting with 3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3 not found: ID does not exist" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.660084 4868 scope.go:117] "RemoveContainer" containerID="eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b" Dec 01 17:45:35 crc kubenswrapper[4868]: E1201 17:45:35.661155 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b\": container with ID starting with eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b not found: ID does not exist" containerID="eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.661180 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b"} err="failed to get container status \"eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b\": rpc error: code = NotFound desc = could not find container \"eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b\": container with ID starting with eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b not found: ID does not exist" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.661194 4868 scope.go:117] "RemoveContainer" containerID="3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.664409 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 17:45:35 crc kubenswrapper[4868]: E1201 17:45:35.664825 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb439a97-8909-414a-a4cf-22e94e909356" containerName="nova-manage" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.664844 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb439a97-8909-414a-a4cf-22e94e909356" containerName="nova-manage" Dec 01 17:45:35 crc kubenswrapper[4868]: E1201 17:45:35.664859 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" containerName="nova-metadata-metadata" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.664866 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" containerName="nova-metadata-metadata" Dec 01 17:45:35 crc kubenswrapper[4868]: E1201 17:45:35.664878 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69c7a38b-c21b-46cd-a120-1d4cd1867941" containerName="dnsmasq-dns" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.664884 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="69c7a38b-c21b-46cd-a120-1d4cd1867941" containerName="dnsmasq-dns" Dec 01 17:45:35 crc kubenswrapper[4868]: E1201 17:45:35.664907 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="897da32f-2dbb-416b-b58e-a4d59040c89c" containerName="nova-cell1-conductor-db-sync" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.664913 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="897da32f-2dbb-416b-b58e-a4d59040c89c" containerName="nova-cell1-conductor-db-sync" Dec 01 17:45:35 crc kubenswrapper[4868]: E1201 17:45:35.664923 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69c7a38b-c21b-46cd-a120-1d4cd1867941" containerName="init" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.664929 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="69c7a38b-c21b-46cd-a120-1d4cd1867941" containerName="init" Dec 01 17:45:35 crc kubenswrapper[4868]: E1201 17:45:35.664964 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" containerName="nova-metadata-log" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.664972 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" containerName="nova-metadata-log" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.665028 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3"} err="failed to get container status \"3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3\": rpc error: code = NotFound desc = could not find container \"3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3\": container with ID starting with 3396ef6fe8f12d2c86608eb49245065501c313a15c6f6d3669f324b1303be8d3 not found: ID does not exist" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.665049 4868 scope.go:117] "RemoveContainer" containerID="eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.665144 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" containerName="nova-metadata-log" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.665156 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="69c7a38b-c21b-46cd-a120-1d4cd1867941" containerName="dnsmasq-dns" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.665166 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" containerName="nova-metadata-metadata" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.665183 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="897da32f-2dbb-416b-b58e-a4d59040c89c" containerName="nova-cell1-conductor-db-sync" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.665192 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb439a97-8909-414a-a4cf-22e94e909356" containerName="nova-manage" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.665789 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.666371 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b"} err="failed to get container status \"eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b\": rpc error: code = NotFound desc = could not find container \"eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b\": container with ID starting with eacac8135a468801043d9449fcb8568342c8000a4c2c07ea19cd96d4524d828b not found: ID does not exist" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.672519 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.677930 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.688098 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.689673 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.693659 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.697515 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.699771 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.812586 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-config-data\") pod \"nova-metadata-0\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.812814 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.812907 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.813009 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.813132 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kv8q\" (UniqueName: \"kubernetes.io/projected/56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c-kube-api-access-8kv8q\") pod \"nova-cell1-conductor-0\" (UID: \"56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.813214 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3239cfe0-6ea9-4580-b350-0ebeb56e2709-logs\") pod \"nova-metadata-0\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.813296 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.813435 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q75c\" (UniqueName: \"kubernetes.io/projected/3239cfe0-6ea9-4580-b350-0ebeb56e2709-kube-api-access-2q75c\") pod \"nova-metadata-0\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.914720 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q75c\" (UniqueName: \"kubernetes.io/projected/3239cfe0-6ea9-4580-b350-0ebeb56e2709-kube-api-access-2q75c\") pod \"nova-metadata-0\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.915005 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-config-data\") pod \"nova-metadata-0\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.915124 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.915207 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.915728 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.915840 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kv8q\" (UniqueName: \"kubernetes.io/projected/56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c-kube-api-access-8kv8q\") pod \"nova-cell1-conductor-0\" (UID: \"56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.915925 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3239cfe0-6ea9-4580-b350-0ebeb56e2709-logs\") pod \"nova-metadata-0\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.916049 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.918504 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3239cfe0-6ea9-4580-b350-0ebeb56e2709-logs\") pod \"nova-metadata-0\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.920518 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.921390 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.921796 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-config-data\") pod \"nova-metadata-0\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.921925 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.924631 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.936442 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kv8q\" (UniqueName: \"kubernetes.io/projected/56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c-kube-api-access-8kv8q\") pod \"nova-cell1-conductor-0\" (UID: \"56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.937849 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q75c\" (UniqueName: \"kubernetes.io/projected/3239cfe0-6ea9-4580-b350-0ebeb56e2709-kube-api-access-2q75c\") pod \"nova-metadata-0\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " pod="openstack/nova-metadata-0" Dec 01 17:45:35 crc kubenswrapper[4868]: I1201 17:45:35.983304 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 17:45:36 crc kubenswrapper[4868]: I1201 17:45:36.010330 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:45:36 crc kubenswrapper[4868]: I1201 17:45:36.193388 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69c7a38b-c21b-46cd-a120-1d4cd1867941" path="/var/lib/kubelet/pods/69c7a38b-c21b-46cd-a120-1d4cd1867941/volumes" Dec 01 17:45:36 crc kubenswrapper[4868]: I1201 17:45:36.194529 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fb1678e-e191-43d1-a9a9-d50b83dfe8f7" path="/var/lib/kubelet/pods/9fb1678e-e191-43d1-a9a9-d50b83dfe8f7/volumes" Dec 01 17:45:36 crc kubenswrapper[4868]: I1201 17:45:36.432833 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 17:45:36 crc kubenswrapper[4868]: I1201 17:45:36.534857 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c","Type":"ContainerStarted","Data":"e866672fe0d3c443d973481def71585f147e2dd646117cce93479ffa3b5f6725"} Dec 01 17:45:36 crc kubenswrapper[4868]: I1201 17:45:36.544815 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ba81079d-403c-45fa-b29d-df13d2d927d8" containerName="nova-scheduler-scheduler" containerID="cri-o://2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f" gracePeriod=30 Dec 01 17:45:36 crc kubenswrapper[4868]: I1201 17:45:36.552447 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:45:36 crc kubenswrapper[4868]: W1201 17:45:36.556777 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3239cfe0_6ea9_4580_b350_0ebeb56e2709.slice/crio-327f2cd2dbabbd3643cfd0ce04fe58d6e901b7439d55718561c251d7b0dc8017 WatchSource:0}: Error finding container 327f2cd2dbabbd3643cfd0ce04fe58d6e901b7439d55718561c251d7b0dc8017: Status 404 returned error can't find the container with id 327f2cd2dbabbd3643cfd0ce04fe58d6e901b7439d55718561c251d7b0dc8017 Dec 01 17:45:37 crc kubenswrapper[4868]: I1201 17:45:37.565012 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c","Type":"ContainerStarted","Data":"45c4b172c9b80bc2a1cce3ad067d207cb6c398cd01439b7a34a0c0279cb5b5e5"} Dec 01 17:45:37 crc kubenswrapper[4868]: I1201 17:45:37.565614 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 01 17:45:37 crc kubenswrapper[4868]: I1201 17:45:37.567024 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3239cfe0-6ea9-4580-b350-0ebeb56e2709","Type":"ContainerStarted","Data":"27b00cadf31aa7b4fe177daddbe129314acbd395f8ed1c08a61339775c6bd911"} Dec 01 17:45:37 crc kubenswrapper[4868]: I1201 17:45:37.567074 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3239cfe0-6ea9-4580-b350-0ebeb56e2709","Type":"ContainerStarted","Data":"a986a0fbad2a9997257c0141ad8464f3d6acf24543c9f22bda40ef0b0ee579bc"} Dec 01 17:45:37 crc kubenswrapper[4868]: I1201 17:45:37.567093 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3239cfe0-6ea9-4580-b350-0ebeb56e2709","Type":"ContainerStarted","Data":"327f2cd2dbabbd3643cfd0ce04fe58d6e901b7439d55718561c251d7b0dc8017"} Dec 01 17:45:37 crc kubenswrapper[4868]: I1201 17:45:37.620460 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.620432541 podStartE2EDuration="2.620432541s" podCreationTimestamp="2025-12-01 17:45:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:45:37.584756403 +0000 UTC m=+1209.955866824" watchObservedRunningTime="2025-12-01 17:45:37.620432541 +0000 UTC m=+1209.991542972" Dec 01 17:45:37 crc kubenswrapper[4868]: I1201 17:45:37.634370 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.634347953 podStartE2EDuration="2.634347953s" podCreationTimestamp="2025-12-01 17:45:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:45:37.607751624 +0000 UTC m=+1209.978862045" watchObservedRunningTime="2025-12-01 17:45:37.634347953 +0000 UTC m=+1210.005458364" Dec 01 17:45:38 crc kubenswrapper[4868]: I1201 17:45:38.734532 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 17:45:38 crc kubenswrapper[4868]: I1201 17:45:38.734762 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="d5daf33c-c3fa-4f86-a422-93e7afa1afaa" containerName="kube-state-metrics" containerID="cri-o://dcb65394c030dbca7a01e84645d6f658c59fe19686951b6c41259c6b905ca678" gracePeriod=30 Dec 01 17:45:38 crc kubenswrapper[4868]: E1201 17:45:38.735596 4868 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 17:45:38 crc kubenswrapper[4868]: E1201 17:45:38.739997 4868 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 17:45:38 crc kubenswrapper[4868]: E1201 17:45:38.741631 4868 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 17:45:38 crc kubenswrapper[4868]: E1201 17:45:38.741715 4868 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ba81079d-403c-45fa-b29d-df13d2d927d8" containerName="nova-scheduler-scheduler" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.209166 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.289147 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kktwt\" (UniqueName: \"kubernetes.io/projected/d5daf33c-c3fa-4f86-a422-93e7afa1afaa-kube-api-access-kktwt\") pod \"d5daf33c-c3fa-4f86-a422-93e7afa1afaa\" (UID: \"d5daf33c-c3fa-4f86-a422-93e7afa1afaa\") " Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.304398 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5daf33c-c3fa-4f86-a422-93e7afa1afaa-kube-api-access-kktwt" (OuterVolumeSpecName: "kube-api-access-kktwt") pod "d5daf33c-c3fa-4f86-a422-93e7afa1afaa" (UID: "d5daf33c-c3fa-4f86-a422-93e7afa1afaa"). InnerVolumeSpecName "kube-api-access-kktwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.392074 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kktwt\" (UniqueName: \"kubernetes.io/projected/d5daf33c-c3fa-4f86-a422-93e7afa1afaa-kube-api-access-kktwt\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.612300 4868 generic.go:334] "Generic (PLEG): container finished" podID="d5daf33c-c3fa-4f86-a422-93e7afa1afaa" containerID="dcb65394c030dbca7a01e84645d6f658c59fe19686951b6c41259c6b905ca678" exitCode=2 Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.612373 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.612393 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d5daf33c-c3fa-4f86-a422-93e7afa1afaa","Type":"ContainerDied","Data":"dcb65394c030dbca7a01e84645d6f658c59fe19686951b6c41259c6b905ca678"} Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.612724 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d5daf33c-c3fa-4f86-a422-93e7afa1afaa","Type":"ContainerDied","Data":"f11a96e3bc572781b287a40771b84921ac365c5adfc5c5ae04545ca0a5395442"} Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.612741 4868 scope.go:117] "RemoveContainer" containerID="dcb65394c030dbca7a01e84645d6f658c59fe19686951b6c41259c6b905ca678" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.640452 4868 scope.go:117] "RemoveContainer" containerID="dcb65394c030dbca7a01e84645d6f658c59fe19686951b6c41259c6b905ca678" Dec 01 17:45:39 crc kubenswrapper[4868]: E1201 17:45:39.640857 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcb65394c030dbca7a01e84645d6f658c59fe19686951b6c41259c6b905ca678\": container with ID starting with dcb65394c030dbca7a01e84645d6f658c59fe19686951b6c41259c6b905ca678 not found: ID does not exist" containerID="dcb65394c030dbca7a01e84645d6f658c59fe19686951b6c41259c6b905ca678" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.640911 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcb65394c030dbca7a01e84645d6f658c59fe19686951b6c41259c6b905ca678"} err="failed to get container status \"dcb65394c030dbca7a01e84645d6f658c59fe19686951b6c41259c6b905ca678\": rpc error: code = NotFound desc = could not find container \"dcb65394c030dbca7a01e84645d6f658c59fe19686951b6c41259c6b905ca678\": container with ID starting with dcb65394c030dbca7a01e84645d6f658c59fe19686951b6c41259c6b905ca678 not found: ID does not exist" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.647096 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.661455 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.671376 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 17:45:39 crc kubenswrapper[4868]: E1201 17:45:39.671893 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5daf33c-c3fa-4f86-a422-93e7afa1afaa" containerName="kube-state-metrics" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.671919 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5daf33c-c3fa-4f86-a422-93e7afa1afaa" containerName="kube-state-metrics" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.672168 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5daf33c-c3fa-4f86-a422-93e7afa1afaa" containerName="kube-state-metrics" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.672776 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.675846 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.676104 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.682869 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.802404 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/36fb76da-24cc-4d21-898e-107fd20eb6a0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"36fb76da-24cc-4d21-898e-107fd20eb6a0\") " pod="openstack/kube-state-metrics-0" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.802483 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36fb76da-24cc-4d21-898e-107fd20eb6a0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"36fb76da-24cc-4d21-898e-107fd20eb6a0\") " pod="openstack/kube-state-metrics-0" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.802734 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/36fb76da-24cc-4d21-898e-107fd20eb6a0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"36fb76da-24cc-4d21-898e-107fd20eb6a0\") " pod="openstack/kube-state-metrics-0" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.802883 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjslj\" (UniqueName: \"kubernetes.io/projected/36fb76da-24cc-4d21-898e-107fd20eb6a0-kube-api-access-wjslj\") pod \"kube-state-metrics-0\" (UID: \"36fb76da-24cc-4d21-898e-107fd20eb6a0\") " pod="openstack/kube-state-metrics-0" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.904816 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjslj\" (UniqueName: \"kubernetes.io/projected/36fb76da-24cc-4d21-898e-107fd20eb6a0-kube-api-access-wjslj\") pod \"kube-state-metrics-0\" (UID: \"36fb76da-24cc-4d21-898e-107fd20eb6a0\") " pod="openstack/kube-state-metrics-0" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.904960 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/36fb76da-24cc-4d21-898e-107fd20eb6a0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"36fb76da-24cc-4d21-898e-107fd20eb6a0\") " pod="openstack/kube-state-metrics-0" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.905010 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36fb76da-24cc-4d21-898e-107fd20eb6a0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"36fb76da-24cc-4d21-898e-107fd20eb6a0\") " pod="openstack/kube-state-metrics-0" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.905077 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/36fb76da-24cc-4d21-898e-107fd20eb6a0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"36fb76da-24cc-4d21-898e-107fd20eb6a0\") " pod="openstack/kube-state-metrics-0" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.910930 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/36fb76da-24cc-4d21-898e-107fd20eb6a0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"36fb76da-24cc-4d21-898e-107fd20eb6a0\") " pod="openstack/kube-state-metrics-0" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.915643 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/36fb76da-24cc-4d21-898e-107fd20eb6a0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"36fb76da-24cc-4d21-898e-107fd20eb6a0\") " pod="openstack/kube-state-metrics-0" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.916572 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36fb76da-24cc-4d21-898e-107fd20eb6a0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"36fb76da-24cc-4d21-898e-107fd20eb6a0\") " pod="openstack/kube-state-metrics-0" Dec 01 17:45:39 crc kubenswrapper[4868]: I1201 17:45:39.927142 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjslj\" (UniqueName: \"kubernetes.io/projected/36fb76da-24cc-4d21-898e-107fd20eb6a0-kube-api-access-wjslj\") pod \"kube-state-metrics-0\" (UID: \"36fb76da-24cc-4d21-898e-107fd20eb6a0\") " pod="openstack/kube-state-metrics-0" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.004043 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.194260 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5daf33c-c3fa-4f86-a422-93e7afa1afaa" path="/var/lib/kubelet/pods/d5daf33c-c3fa-4f86-a422-93e7afa1afaa/volumes" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.440628 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.465539 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.580441 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.622499 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"36fb76da-24cc-4d21-898e-107fd20eb6a0","Type":"ContainerStarted","Data":"e09ac0f4fd5fc457f8958b63f61284ae3af7a15ce2772aceb1ba3342eab3e45c"} Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.624191 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6v57\" (UniqueName: \"kubernetes.io/projected/ba81079d-403c-45fa-b29d-df13d2d927d8-kube-api-access-k6v57\") pod \"ba81079d-403c-45fa-b29d-df13d2d927d8\" (UID: \"ba81079d-403c-45fa-b29d-df13d2d927d8\") " Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.624274 4868 generic.go:334] "Generic (PLEG): container finished" podID="a1bc0fc1-0afa-4e05-abe0-05285c85576f" containerID="e84d1798bc0ae4641d48befea34912af24a042cf4624eb73c2fe18000183365e" exitCode=0 Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.624324 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba81079d-403c-45fa-b29d-df13d2d927d8-config-data\") pod \"ba81079d-403c-45fa-b29d-df13d2d927d8\" (UID: \"ba81079d-403c-45fa-b29d-df13d2d927d8\") " Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.624358 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.624374 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba81079d-403c-45fa-b29d-df13d2d927d8-combined-ca-bundle\") pod \"ba81079d-403c-45fa-b29d-df13d2d927d8\" (UID: \"ba81079d-403c-45fa-b29d-df13d2d927d8\") " Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.624534 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a1bc0fc1-0afa-4e05-abe0-05285c85576f","Type":"ContainerDied","Data":"e84d1798bc0ae4641d48befea34912af24a042cf4624eb73c2fe18000183365e"} Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.624566 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a1bc0fc1-0afa-4e05-abe0-05285c85576f","Type":"ContainerDied","Data":"2246a3aafd627aa81555347369cc896943681daa6266504618279404b5a000e8"} Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.624586 4868 scope.go:117] "RemoveContainer" containerID="e84d1798bc0ae4641d48befea34912af24a042cf4624eb73c2fe18000183365e" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.630865 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba81079d-403c-45fa-b29d-df13d2d927d8-kube-api-access-k6v57" (OuterVolumeSpecName: "kube-api-access-k6v57") pod "ba81079d-403c-45fa-b29d-df13d2d927d8" (UID: "ba81079d-403c-45fa-b29d-df13d2d927d8"). InnerVolumeSpecName "kube-api-access-k6v57". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.636817 4868 generic.go:334] "Generic (PLEG): container finished" podID="ba81079d-403c-45fa-b29d-df13d2d927d8" containerID="2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f" exitCode=0 Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.636925 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.637295 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ba81079d-403c-45fa-b29d-df13d2d927d8","Type":"ContainerDied","Data":"2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f"} Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.644612 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ba81079d-403c-45fa-b29d-df13d2d927d8","Type":"ContainerDied","Data":"066ff1affb5f83b9e7743ae4e41783ff4178240facb8ad9307ba3309fd76c145"} Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.670049 4868 scope.go:117] "RemoveContainer" containerID="581755dab82fc21797870f98cdbd213bc13b37d56d056a1c91852285b1b4e4db" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.674140 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba81079d-403c-45fa-b29d-df13d2d927d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba81079d-403c-45fa-b29d-df13d2d927d8" (UID: "ba81079d-403c-45fa-b29d-df13d2d927d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.676625 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba81079d-403c-45fa-b29d-df13d2d927d8-config-data" (OuterVolumeSpecName: "config-data") pod "ba81079d-403c-45fa-b29d-df13d2d927d8" (UID: "ba81079d-403c-45fa-b29d-df13d2d927d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.692151 4868 scope.go:117] "RemoveContainer" containerID="e84d1798bc0ae4641d48befea34912af24a042cf4624eb73c2fe18000183365e" Dec 01 17:45:40 crc kubenswrapper[4868]: E1201 17:45:40.692555 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e84d1798bc0ae4641d48befea34912af24a042cf4624eb73c2fe18000183365e\": container with ID starting with e84d1798bc0ae4641d48befea34912af24a042cf4624eb73c2fe18000183365e not found: ID does not exist" containerID="e84d1798bc0ae4641d48befea34912af24a042cf4624eb73c2fe18000183365e" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.692588 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e84d1798bc0ae4641d48befea34912af24a042cf4624eb73c2fe18000183365e"} err="failed to get container status \"e84d1798bc0ae4641d48befea34912af24a042cf4624eb73c2fe18000183365e\": rpc error: code = NotFound desc = could not find container \"e84d1798bc0ae4641d48befea34912af24a042cf4624eb73c2fe18000183365e\": container with ID starting with e84d1798bc0ae4641d48befea34912af24a042cf4624eb73c2fe18000183365e not found: ID does not exist" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.692608 4868 scope.go:117] "RemoveContainer" containerID="581755dab82fc21797870f98cdbd213bc13b37d56d056a1c91852285b1b4e4db" Dec 01 17:45:40 crc kubenswrapper[4868]: E1201 17:45:40.692985 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"581755dab82fc21797870f98cdbd213bc13b37d56d056a1c91852285b1b4e4db\": container with ID starting with 581755dab82fc21797870f98cdbd213bc13b37d56d056a1c91852285b1b4e4db not found: ID does not exist" containerID="581755dab82fc21797870f98cdbd213bc13b37d56d056a1c91852285b1b4e4db" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.693026 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"581755dab82fc21797870f98cdbd213bc13b37d56d056a1c91852285b1b4e4db"} err="failed to get container status \"581755dab82fc21797870f98cdbd213bc13b37d56d056a1c91852285b1b4e4db\": rpc error: code = NotFound desc = could not find container \"581755dab82fc21797870f98cdbd213bc13b37d56d056a1c91852285b1b4e4db\": container with ID starting with 581755dab82fc21797870f98cdbd213bc13b37d56d056a1c91852285b1b4e4db not found: ID does not exist" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.693043 4868 scope.go:117] "RemoveContainer" containerID="2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.728420 4868 scope.go:117] "RemoveContainer" containerID="2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f" Dec 01 17:45:40 crc kubenswrapper[4868]: E1201 17:45:40.728913 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f\": container with ID starting with 2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f not found: ID does not exist" containerID="2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.728959 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f"} err="failed to get container status \"2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f\": rpc error: code = NotFound desc = could not find container \"2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f\": container with ID starting with 2bc51b7e178b668afdc9f9cdee1069276a6b73664b10f4ee3c690c9d1b1c6e3f not found: ID does not exist" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.729269 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1bc0fc1-0afa-4e05-abe0-05285c85576f-combined-ca-bundle\") pod \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.729354 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1bc0fc1-0afa-4e05-abe0-05285c85576f-config-data\") pod \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.729447 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgdgv\" (UniqueName: \"kubernetes.io/projected/a1bc0fc1-0afa-4e05-abe0-05285c85576f-kube-api-access-pgdgv\") pod \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.729489 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1bc0fc1-0afa-4e05-abe0-05285c85576f-logs\") pod \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.730002 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6v57\" (UniqueName: \"kubernetes.io/projected/ba81079d-403c-45fa-b29d-df13d2d927d8-kube-api-access-k6v57\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.730028 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba81079d-403c-45fa-b29d-df13d2d927d8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.730038 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba81079d-403c-45fa-b29d-df13d2d927d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.730918 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1bc0fc1-0afa-4e05-abe0-05285c85576f-logs" (OuterVolumeSpecName: "logs") pod "a1bc0fc1-0afa-4e05-abe0-05285c85576f" (UID: "a1bc0fc1-0afa-4e05-abe0-05285c85576f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.737195 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1bc0fc1-0afa-4e05-abe0-05285c85576f-kube-api-access-pgdgv" (OuterVolumeSpecName: "kube-api-access-pgdgv") pod "a1bc0fc1-0afa-4e05-abe0-05285c85576f" (UID: "a1bc0fc1-0afa-4e05-abe0-05285c85576f"). InnerVolumeSpecName "kube-api-access-pgdgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.760900 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.761195 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="ceilometer-central-agent" containerID="cri-o://26ba2d03acdd9505144fa1012002e04da97c9687cdffc890667f92030f2421b0" gracePeriod=30 Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.761322 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="ceilometer-notification-agent" containerID="cri-o://dfdeae37b27c6cdc9c4b5428bb5dd3a1a91f78c21054d7315effb2577bf49b91" gracePeriod=30 Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.761301 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="sg-core" containerID="cri-o://32aae52f5b745b711e8c82229ce99775a3b15bb0c9dbb65806e5a29ed6efad33" gracePeriod=30 Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.761459 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="proxy-httpd" containerID="cri-o://90efad9f1a86acadc5d0089dd3923d5a9bff46b9f8cd7e5b7bc64bf9d023349f" gracePeriod=30 Dec 01 17:45:40 crc kubenswrapper[4868]: E1201 17:45:40.778105 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a1bc0fc1-0afa-4e05-abe0-05285c85576f-combined-ca-bundle podName:a1bc0fc1-0afa-4e05-abe0-05285c85576f nodeName:}" failed. No retries permitted until 2025-12-01 17:45:41.278074279 +0000 UTC m=+1213.649184690 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/a1bc0fc1-0afa-4e05-abe0-05285c85576f-combined-ca-bundle") pod "a1bc0fc1-0afa-4e05-abe0-05285c85576f" (UID: "a1bc0fc1-0afa-4e05-abe0-05285c85576f") : error deleting /var/lib/kubelet/pods/a1bc0fc1-0afa-4e05-abe0-05285c85576f/volume-subpaths: remove /var/lib/kubelet/pods/a1bc0fc1-0afa-4e05-abe0-05285c85576f/volume-subpaths: no such file or directory Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.781199 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1bc0fc1-0afa-4e05-abe0-05285c85576f-config-data" (OuterVolumeSpecName: "config-data") pod "a1bc0fc1-0afa-4e05-abe0-05285c85576f" (UID: "a1bc0fc1-0afa-4e05-abe0-05285c85576f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.831750 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1bc0fc1-0afa-4e05-abe0-05285c85576f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.832880 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgdgv\" (UniqueName: \"kubernetes.io/projected/a1bc0fc1-0afa-4e05-abe0-05285c85576f-kube-api-access-pgdgv\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:40 crc kubenswrapper[4868]: I1201 17:45:40.832960 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1bc0fc1-0afa-4e05-abe0-05285c85576f-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.011349 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.011408 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.064335 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.073647 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.084320 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:45:41 crc kubenswrapper[4868]: E1201 17:45:41.085124 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1bc0fc1-0afa-4e05-abe0-05285c85576f" containerName="nova-api-log" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.085144 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1bc0fc1-0afa-4e05-abe0-05285c85576f" containerName="nova-api-log" Dec 01 17:45:41 crc kubenswrapper[4868]: E1201 17:45:41.085158 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1bc0fc1-0afa-4e05-abe0-05285c85576f" containerName="nova-api-api" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.085164 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1bc0fc1-0afa-4e05-abe0-05285c85576f" containerName="nova-api-api" Dec 01 17:45:41 crc kubenswrapper[4868]: E1201 17:45:41.085188 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba81079d-403c-45fa-b29d-df13d2d927d8" containerName="nova-scheduler-scheduler" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.085194 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba81079d-403c-45fa-b29d-df13d2d927d8" containerName="nova-scheduler-scheduler" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.085355 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1bc0fc1-0afa-4e05-abe0-05285c85576f" containerName="nova-api-api" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.085383 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba81079d-403c-45fa-b29d-df13d2d927d8" containerName="nova-scheduler-scheduler" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.085391 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1bc0fc1-0afa-4e05-abe0-05285c85576f" containerName="nova-api-log" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.086092 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.088695 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.092265 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.240208 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgvn5\" (UniqueName: \"kubernetes.io/projected/0eb5aa87-3260-4580-8827-b6a740ff01eb-kube-api-access-qgvn5\") pod \"nova-scheduler-0\" (UID: \"0eb5aa87-3260-4580-8827-b6a740ff01eb\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.240302 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0eb5aa87-3260-4580-8827-b6a740ff01eb-config-data\") pod \"nova-scheduler-0\" (UID: \"0eb5aa87-3260-4580-8827-b6a740ff01eb\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.240452 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0eb5aa87-3260-4580-8827-b6a740ff01eb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0eb5aa87-3260-4580-8827-b6a740ff01eb\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.342220 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1bc0fc1-0afa-4e05-abe0-05285c85576f-combined-ca-bundle\") pod \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\" (UID: \"a1bc0fc1-0afa-4e05-abe0-05285c85576f\") " Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.343048 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgvn5\" (UniqueName: \"kubernetes.io/projected/0eb5aa87-3260-4580-8827-b6a740ff01eb-kube-api-access-qgvn5\") pod \"nova-scheduler-0\" (UID: \"0eb5aa87-3260-4580-8827-b6a740ff01eb\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.344000 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0eb5aa87-3260-4580-8827-b6a740ff01eb-config-data\") pod \"nova-scheduler-0\" (UID: \"0eb5aa87-3260-4580-8827-b6a740ff01eb\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.344202 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0eb5aa87-3260-4580-8827-b6a740ff01eb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0eb5aa87-3260-4580-8827-b6a740ff01eb\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.347402 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1bc0fc1-0afa-4e05-abe0-05285c85576f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1bc0fc1-0afa-4e05-abe0-05285c85576f" (UID: "a1bc0fc1-0afa-4e05-abe0-05285c85576f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.349602 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0eb5aa87-3260-4580-8827-b6a740ff01eb-config-data\") pod \"nova-scheduler-0\" (UID: \"0eb5aa87-3260-4580-8827-b6a740ff01eb\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.354557 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0eb5aa87-3260-4580-8827-b6a740ff01eb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0eb5aa87-3260-4580-8827-b6a740ff01eb\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.361394 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgvn5\" (UniqueName: \"kubernetes.io/projected/0eb5aa87-3260-4580-8827-b6a740ff01eb-kube-api-access-qgvn5\") pod \"nova-scheduler-0\" (UID: \"0eb5aa87-3260-4580-8827-b6a740ff01eb\") " pod="openstack/nova-scheduler-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.407578 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.446886 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1bc0fc1-0afa-4e05-abe0-05285c85576f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.592620 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.604203 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.616969 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.618623 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.621156 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.639882 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.666647 4868 generic.go:334] "Generic (PLEG): container finished" podID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerID="90efad9f1a86acadc5d0089dd3923d5a9bff46b9f8cd7e5b7bc64bf9d023349f" exitCode=0 Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.666687 4868 generic.go:334] "Generic (PLEG): container finished" podID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerID="32aae52f5b745b711e8c82229ce99775a3b15bb0c9dbb65806e5a29ed6efad33" exitCode=2 Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.666698 4868 generic.go:334] "Generic (PLEG): container finished" podID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerID="26ba2d03acdd9505144fa1012002e04da97c9687cdffc890667f92030f2421b0" exitCode=0 Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.666714 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cda1e79-77f9-4b20-a0cf-cf8f0619228b","Type":"ContainerDied","Data":"90efad9f1a86acadc5d0089dd3923d5a9bff46b9f8cd7e5b7bc64bf9d023349f"} Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.666766 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cda1e79-77f9-4b20-a0cf-cf8f0619228b","Type":"ContainerDied","Data":"32aae52f5b745b711e8c82229ce99775a3b15bb0c9dbb65806e5a29ed6efad33"} Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.666784 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cda1e79-77f9-4b20-a0cf-cf8f0619228b","Type":"ContainerDied","Data":"26ba2d03acdd9505144fa1012002e04da97c9687cdffc890667f92030f2421b0"} Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.669070 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"36fb76da-24cc-4d21-898e-107fd20eb6a0","Type":"ContainerStarted","Data":"3a6616e8e8330c0a29359ddd8ca8e25f52ee431282ef3fbd94cea1587328c7e1"} Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.669339 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.691336 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.319943118 podStartE2EDuration="2.691320457s" podCreationTimestamp="2025-12-01 17:45:39 +0000 UTC" firstStartedPulling="2025-12-01 17:45:40.496993066 +0000 UTC m=+1212.868103467" lastFinishedPulling="2025-12-01 17:45:40.868370395 +0000 UTC m=+1213.239480806" observedRunningTime="2025-12-01 17:45:41.687144204 +0000 UTC m=+1214.058254615" watchObservedRunningTime="2025-12-01 17:45:41.691320457 +0000 UTC m=+1214.062430868" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.762609 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvp7s\" (UniqueName: \"kubernetes.io/projected/430e5d96-e42d-433f-b311-39a7ab09a98c-kube-api-access-nvp7s\") pod \"nova-api-0\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " pod="openstack/nova-api-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.762841 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430e5d96-e42d-433f-b311-39a7ab09a98c-config-data\") pod \"nova-api-0\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " pod="openstack/nova-api-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.763077 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430e5d96-e42d-433f-b311-39a7ab09a98c-logs\") pod \"nova-api-0\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " pod="openstack/nova-api-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.763456 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430e5d96-e42d-433f-b311-39a7ab09a98c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " pod="openstack/nova-api-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.864979 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvp7s\" (UniqueName: \"kubernetes.io/projected/430e5d96-e42d-433f-b311-39a7ab09a98c-kube-api-access-nvp7s\") pod \"nova-api-0\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " pod="openstack/nova-api-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.865046 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430e5d96-e42d-433f-b311-39a7ab09a98c-config-data\") pod \"nova-api-0\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " pod="openstack/nova-api-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.865083 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430e5d96-e42d-433f-b311-39a7ab09a98c-logs\") pod \"nova-api-0\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " pod="openstack/nova-api-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.865163 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430e5d96-e42d-433f-b311-39a7ab09a98c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " pod="openstack/nova-api-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.865704 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430e5d96-e42d-433f-b311-39a7ab09a98c-logs\") pod \"nova-api-0\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " pod="openstack/nova-api-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.868690 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430e5d96-e42d-433f-b311-39a7ab09a98c-config-data\") pod \"nova-api-0\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " pod="openstack/nova-api-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.868900 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430e5d96-e42d-433f-b311-39a7ab09a98c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " pod="openstack/nova-api-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.880230 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvp7s\" (UniqueName: \"kubernetes.io/projected/430e5d96-e42d-433f-b311-39a7ab09a98c-kube-api-access-nvp7s\") pod \"nova-api-0\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " pod="openstack/nova-api-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.942871 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:45:41 crc kubenswrapper[4868]: I1201 17:45:41.943034 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:45:42 crc kubenswrapper[4868]: I1201 17:45:42.202484 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1bc0fc1-0afa-4e05-abe0-05285c85576f" path="/var/lib/kubelet/pods/a1bc0fc1-0afa-4e05-abe0-05285c85576f/volumes" Dec 01 17:45:42 crc kubenswrapper[4868]: I1201 17:45:42.203114 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba81079d-403c-45fa-b29d-df13d2d927d8" path="/var/lib/kubelet/pods/ba81079d-403c-45fa-b29d-df13d2d927d8/volumes" Dec 01 17:45:42 crc kubenswrapper[4868]: I1201 17:45:42.383246 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:45:42 crc kubenswrapper[4868]: W1201 17:45:42.387817 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod430e5d96_e42d_433f_b311_39a7ab09a98c.slice/crio-5e1f593e47118489217ee5e49cc3a196e5725bc8a40a66ef069a19e6f99ea093 WatchSource:0}: Error finding container 5e1f593e47118489217ee5e49cc3a196e5725bc8a40a66ef069a19e6f99ea093: Status 404 returned error can't find the container with id 5e1f593e47118489217ee5e49cc3a196e5725bc8a40a66ef069a19e6f99ea093 Dec 01 17:45:42 crc kubenswrapper[4868]: I1201 17:45:42.681604 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"430e5d96-e42d-433f-b311-39a7ab09a98c","Type":"ContainerStarted","Data":"433ab3fd5801bc8a68904a86200c42f15979352c6babf8f6fc1cf60ebf1fc7bf"} Dec 01 17:45:42 crc kubenswrapper[4868]: I1201 17:45:42.681649 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"430e5d96-e42d-433f-b311-39a7ab09a98c","Type":"ContainerStarted","Data":"5e1f593e47118489217ee5e49cc3a196e5725bc8a40a66ef069a19e6f99ea093"} Dec 01 17:45:42 crc kubenswrapper[4868]: I1201 17:45:42.684154 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0eb5aa87-3260-4580-8827-b6a740ff01eb","Type":"ContainerStarted","Data":"ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482"} Dec 01 17:45:42 crc kubenswrapper[4868]: I1201 17:45:42.684240 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0eb5aa87-3260-4580-8827-b6a740ff01eb","Type":"ContainerStarted","Data":"212689681e76de1303085bcd70599b720bf8469f60fed03bdb650e3e0816bfc6"} Dec 01 17:45:42 crc kubenswrapper[4868]: I1201 17:45:42.705116 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.705090191 podStartE2EDuration="1.705090191s" podCreationTimestamp="2025-12-01 17:45:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:45:42.700809764 +0000 UTC m=+1215.071920175" watchObservedRunningTime="2025-12-01 17:45:42.705090191 +0000 UTC m=+1215.076200602" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.261379 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.397783 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-scripts\") pod \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.397986 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-combined-ca-bundle\") pod \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.398180 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdqvf\" (UniqueName: \"kubernetes.io/projected/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-kube-api-access-jdqvf\") pod \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.398244 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-run-httpd\") pod \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.398277 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-sg-core-conf-yaml\") pod \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.398307 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-log-httpd\") pod \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.398392 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-config-data\") pod \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\" (UID: \"2cda1e79-77f9-4b20-a0cf-cf8f0619228b\") " Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.398622 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2cda1e79-77f9-4b20-a0cf-cf8f0619228b" (UID: "2cda1e79-77f9-4b20-a0cf-cf8f0619228b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.398869 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2cda1e79-77f9-4b20-a0cf-cf8f0619228b" (UID: "2cda1e79-77f9-4b20-a0cf-cf8f0619228b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.399236 4868 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.399259 4868 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.405547 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-scripts" (OuterVolumeSpecName: "scripts") pod "2cda1e79-77f9-4b20-a0cf-cf8f0619228b" (UID: "2cda1e79-77f9-4b20-a0cf-cf8f0619228b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.422956 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-kube-api-access-jdqvf" (OuterVolumeSpecName: "kube-api-access-jdqvf") pod "2cda1e79-77f9-4b20-a0cf-cf8f0619228b" (UID: "2cda1e79-77f9-4b20-a0cf-cf8f0619228b"). InnerVolumeSpecName "kube-api-access-jdqvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.441484 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2cda1e79-77f9-4b20-a0cf-cf8f0619228b" (UID: "2cda1e79-77f9-4b20-a0cf-cf8f0619228b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.489014 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2cda1e79-77f9-4b20-a0cf-cf8f0619228b" (UID: "2cda1e79-77f9-4b20-a0cf-cf8f0619228b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.502056 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.502105 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.502122 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdqvf\" (UniqueName: \"kubernetes.io/projected/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-kube-api-access-jdqvf\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.502136 4868 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.523309 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-config-data" (OuterVolumeSpecName: "config-data") pod "2cda1e79-77f9-4b20-a0cf-cf8f0619228b" (UID: "2cda1e79-77f9-4b20-a0cf-cf8f0619228b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.604163 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cda1e79-77f9-4b20-a0cf-cf8f0619228b-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.697612 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"430e5d96-e42d-433f-b311-39a7ab09a98c","Type":"ContainerStarted","Data":"0105f3c449df8f1a4888fdb10c28c6358b67a6322627d5ad73a2e7d1f3b73344"} Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.707288 4868 generic.go:334] "Generic (PLEG): container finished" podID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerID="dfdeae37b27c6cdc9c4b5428bb5dd3a1a91f78c21054d7315effb2577bf49b91" exitCode=0 Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.708239 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.716834 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cda1e79-77f9-4b20-a0cf-cf8f0619228b","Type":"ContainerDied","Data":"dfdeae37b27c6cdc9c4b5428bb5dd3a1a91f78c21054d7315effb2577bf49b91"} Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.716930 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cda1e79-77f9-4b20-a0cf-cf8f0619228b","Type":"ContainerDied","Data":"1ff4d68691ffcdf9c4f6f58546fabc2c5a133b4a74a3eab4af21225187113530"} Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.716976 4868 scope.go:117] "RemoveContainer" containerID="90efad9f1a86acadc5d0089dd3923d5a9bff46b9f8cd7e5b7bc64bf9d023349f" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.747456 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.747430307 podStartE2EDuration="2.747430307s" podCreationTimestamp="2025-12-01 17:45:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:45:43.724499679 +0000 UTC m=+1216.095610090" watchObservedRunningTime="2025-12-01 17:45:43.747430307 +0000 UTC m=+1216.118540718" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.759187 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.766484 4868 scope.go:117] "RemoveContainer" containerID="32aae52f5b745b711e8c82229ce99775a3b15bb0c9dbb65806e5a29ed6efad33" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.769037 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.790181 4868 scope.go:117] "RemoveContainer" containerID="dfdeae37b27c6cdc9c4b5428bb5dd3a1a91f78c21054d7315effb2577bf49b91" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.792664 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:45:43 crc kubenswrapper[4868]: E1201 17:45:43.793720 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="sg-core" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.793744 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="sg-core" Dec 01 17:45:43 crc kubenswrapper[4868]: E1201 17:45:43.793762 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="ceilometer-notification-agent" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.793770 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="ceilometer-notification-agent" Dec 01 17:45:43 crc kubenswrapper[4868]: E1201 17:45:43.793782 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="ceilometer-central-agent" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.793788 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="ceilometer-central-agent" Dec 01 17:45:43 crc kubenswrapper[4868]: E1201 17:45:43.793801 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="proxy-httpd" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.793806 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="proxy-httpd" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.793998 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="ceilometer-central-agent" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.794020 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="sg-core" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.794034 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="proxy-httpd" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.794046 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" containerName="ceilometer-notification-agent" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.796990 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.801148 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.814898 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.815244 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.826990 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.838384 4868 scope.go:117] "RemoveContainer" containerID="26ba2d03acdd9505144fa1012002e04da97c9687cdffc890667f92030f2421b0" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.870363 4868 scope.go:117] "RemoveContainer" containerID="90efad9f1a86acadc5d0089dd3923d5a9bff46b9f8cd7e5b7bc64bf9d023349f" Dec 01 17:45:43 crc kubenswrapper[4868]: E1201 17:45:43.871052 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90efad9f1a86acadc5d0089dd3923d5a9bff46b9f8cd7e5b7bc64bf9d023349f\": container with ID starting with 90efad9f1a86acadc5d0089dd3923d5a9bff46b9f8cd7e5b7bc64bf9d023349f not found: ID does not exist" containerID="90efad9f1a86acadc5d0089dd3923d5a9bff46b9f8cd7e5b7bc64bf9d023349f" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.871099 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90efad9f1a86acadc5d0089dd3923d5a9bff46b9f8cd7e5b7bc64bf9d023349f"} err="failed to get container status \"90efad9f1a86acadc5d0089dd3923d5a9bff46b9f8cd7e5b7bc64bf9d023349f\": rpc error: code = NotFound desc = could not find container \"90efad9f1a86acadc5d0089dd3923d5a9bff46b9f8cd7e5b7bc64bf9d023349f\": container with ID starting with 90efad9f1a86acadc5d0089dd3923d5a9bff46b9f8cd7e5b7bc64bf9d023349f not found: ID does not exist" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.871128 4868 scope.go:117] "RemoveContainer" containerID="32aae52f5b745b711e8c82229ce99775a3b15bb0c9dbb65806e5a29ed6efad33" Dec 01 17:45:43 crc kubenswrapper[4868]: E1201 17:45:43.871439 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32aae52f5b745b711e8c82229ce99775a3b15bb0c9dbb65806e5a29ed6efad33\": container with ID starting with 32aae52f5b745b711e8c82229ce99775a3b15bb0c9dbb65806e5a29ed6efad33 not found: ID does not exist" containerID="32aae52f5b745b711e8c82229ce99775a3b15bb0c9dbb65806e5a29ed6efad33" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.871466 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32aae52f5b745b711e8c82229ce99775a3b15bb0c9dbb65806e5a29ed6efad33"} err="failed to get container status \"32aae52f5b745b711e8c82229ce99775a3b15bb0c9dbb65806e5a29ed6efad33\": rpc error: code = NotFound desc = could not find container \"32aae52f5b745b711e8c82229ce99775a3b15bb0c9dbb65806e5a29ed6efad33\": container with ID starting with 32aae52f5b745b711e8c82229ce99775a3b15bb0c9dbb65806e5a29ed6efad33 not found: ID does not exist" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.871480 4868 scope.go:117] "RemoveContainer" containerID="dfdeae37b27c6cdc9c4b5428bb5dd3a1a91f78c21054d7315effb2577bf49b91" Dec 01 17:45:43 crc kubenswrapper[4868]: E1201 17:45:43.871923 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfdeae37b27c6cdc9c4b5428bb5dd3a1a91f78c21054d7315effb2577bf49b91\": container with ID starting with dfdeae37b27c6cdc9c4b5428bb5dd3a1a91f78c21054d7315effb2577bf49b91 not found: ID does not exist" containerID="dfdeae37b27c6cdc9c4b5428bb5dd3a1a91f78c21054d7315effb2577bf49b91" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.871958 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfdeae37b27c6cdc9c4b5428bb5dd3a1a91f78c21054d7315effb2577bf49b91"} err="failed to get container status \"dfdeae37b27c6cdc9c4b5428bb5dd3a1a91f78c21054d7315effb2577bf49b91\": rpc error: code = NotFound desc = could not find container \"dfdeae37b27c6cdc9c4b5428bb5dd3a1a91f78c21054d7315effb2577bf49b91\": container with ID starting with dfdeae37b27c6cdc9c4b5428bb5dd3a1a91f78c21054d7315effb2577bf49b91 not found: ID does not exist" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.871979 4868 scope.go:117] "RemoveContainer" containerID="26ba2d03acdd9505144fa1012002e04da97c9687cdffc890667f92030f2421b0" Dec 01 17:45:43 crc kubenswrapper[4868]: E1201 17:45:43.872460 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26ba2d03acdd9505144fa1012002e04da97c9687cdffc890667f92030f2421b0\": container with ID starting with 26ba2d03acdd9505144fa1012002e04da97c9687cdffc890667f92030f2421b0 not found: ID does not exist" containerID="26ba2d03acdd9505144fa1012002e04da97c9687cdffc890667f92030f2421b0" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.872483 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26ba2d03acdd9505144fa1012002e04da97c9687cdffc890667f92030f2421b0"} err="failed to get container status \"26ba2d03acdd9505144fa1012002e04da97c9687cdffc890667f92030f2421b0\": rpc error: code = NotFound desc = could not find container \"26ba2d03acdd9505144fa1012002e04da97c9687cdffc890667f92030f2421b0\": container with ID starting with 26ba2d03acdd9505144fa1012002e04da97c9687cdffc890667f92030f2421b0 not found: ID does not exist" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.915147 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-run-httpd\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.915778 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.915828 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f57gs\" (UniqueName: \"kubernetes.io/projected/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-kube-api-access-f57gs\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.915962 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-config-data\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.915995 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.916028 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.916058 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-log-httpd\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:43 crc kubenswrapper[4868]: I1201 17:45:43.916086 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-scripts\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.017749 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.017805 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.017826 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-log-httpd\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.017855 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-scripts\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.017880 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-run-httpd\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.017967 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.018000 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f57gs\" (UniqueName: \"kubernetes.io/projected/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-kube-api-access-f57gs\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.018071 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-config-data\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.018376 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-log-httpd\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.018512 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-run-httpd\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.024522 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.024775 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-config-data\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.026795 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.033019 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.050146 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f57gs\" (UniqueName: \"kubernetes.io/projected/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-kube-api-access-f57gs\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.050685 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-scripts\") pod \"ceilometer-0\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.127596 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.205434 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cda1e79-77f9-4b20-a0cf-cf8f0619228b" path="/var/lib/kubelet/pods/2cda1e79-77f9-4b20-a0cf-cf8f0619228b/volumes" Dec 01 17:45:44 crc kubenswrapper[4868]: W1201 17:45:44.571901 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11c523a5_0e2d_4e46_b2c7_516ab7cf2617.slice/crio-0d9599ab48901d40f5f0c3a01a8ca30e9236b322f60a5febf7ccc2377eb96899 WatchSource:0}: Error finding container 0d9599ab48901d40f5f0c3a01a8ca30e9236b322f60a5febf7ccc2377eb96899: Status 404 returned error can't find the container with id 0d9599ab48901d40f5f0c3a01a8ca30e9236b322f60a5febf7ccc2377eb96899 Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.574550 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:45:44 crc kubenswrapper[4868]: I1201 17:45:44.722261 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c523a5-0e2d-4e46-b2c7-516ab7cf2617","Type":"ContainerStarted","Data":"0d9599ab48901d40f5f0c3a01a8ca30e9236b322f60a5febf7ccc2377eb96899"} Dec 01 17:45:45 crc kubenswrapper[4868]: I1201 17:45:45.742316 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c523a5-0e2d-4e46-b2c7-516ab7cf2617","Type":"ContainerStarted","Data":"5ab0302df9caf88ccebb6696b9836c30e56547f52571bfeb47b37170ede16727"} Dec 01 17:45:46 crc kubenswrapper[4868]: I1201 17:45:46.011836 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 17:45:46 crc kubenswrapper[4868]: I1201 17:45:46.012254 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 17:45:46 crc kubenswrapper[4868]: I1201 17:45:46.015287 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 01 17:45:46 crc kubenswrapper[4868]: I1201 17:45:46.408603 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 17:45:46 crc kubenswrapper[4868]: I1201 17:45:46.769741 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c523a5-0e2d-4e46-b2c7-516ab7cf2617","Type":"ContainerStarted","Data":"c633510a6fb594f0774bc484ba14c1ab5165955d6a447c972ce5854be1a6971f"} Dec 01 17:45:47 crc kubenswrapper[4868]: I1201 17:45:47.028076 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3239cfe0-6ea9-4580-b350-0ebeb56e2709" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 17:45:47 crc kubenswrapper[4868]: I1201 17:45:47.028075 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3239cfe0-6ea9-4580-b350-0ebeb56e2709" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 17:45:47 crc kubenswrapper[4868]: I1201 17:45:47.781235 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c523a5-0e2d-4e46-b2c7-516ab7cf2617","Type":"ContainerStarted","Data":"99a1b6fc34c2e67c253ca05527f7e655c2feb0a0b20a20e4ca6aef7ece34d702"} Dec 01 17:45:48 crc kubenswrapper[4868]: I1201 17:45:48.797988 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c523a5-0e2d-4e46-b2c7-516ab7cf2617","Type":"ContainerStarted","Data":"07f613806cd357e8e4387bd714d3cf2b9ec2c76343ac5d461d72829f4f1dc917"} Dec 01 17:45:48 crc kubenswrapper[4868]: I1201 17:45:48.798474 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 17:45:48 crc kubenswrapper[4868]: I1201 17:45:48.826190 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9589230899999999 podStartE2EDuration="5.826167225s" podCreationTimestamp="2025-12-01 17:45:43 +0000 UTC" firstStartedPulling="2025-12-01 17:45:44.575124511 +0000 UTC m=+1216.946234922" lastFinishedPulling="2025-12-01 17:45:48.442368646 +0000 UTC m=+1220.813479057" observedRunningTime="2025-12-01 17:45:48.817308012 +0000 UTC m=+1221.188418473" watchObservedRunningTime="2025-12-01 17:45:48.826167225 +0000 UTC m=+1221.197277636" Dec 01 17:45:50 crc kubenswrapper[4868]: I1201 17:45:50.014271 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 17:45:51 crc kubenswrapper[4868]: I1201 17:45:51.408971 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 17:45:51 crc kubenswrapper[4868]: I1201 17:45:51.437807 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 17:45:51 crc kubenswrapper[4868]: I1201 17:45:51.849713 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 17:45:51 crc kubenswrapper[4868]: I1201 17:45:51.943496 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 17:45:51 crc kubenswrapper[4868]: I1201 17:45:51.943545 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 17:45:53 crc kubenswrapper[4868]: I1201 17:45:53.027171 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="430e5d96-e42d-433f-b311-39a7ab09a98c" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 17:45:53 crc kubenswrapper[4868]: I1201 17:45:53.027284 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="430e5d96-e42d-433f-b311-39a7ab09a98c" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 17:45:55 crc kubenswrapper[4868]: I1201 17:45:55.905003 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:45:55 crc kubenswrapper[4868]: I1201 17:45:55.905819 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:45:56 crc kubenswrapper[4868]: I1201 17:45:56.019742 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 17:45:56 crc kubenswrapper[4868]: I1201 17:45:56.019804 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 17:45:56 crc kubenswrapper[4868]: I1201 17:45:56.027357 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 17:45:56 crc kubenswrapper[4868]: I1201 17:45:56.027405 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 17:45:59 crc kubenswrapper[4868]: I1201 17:45:59.906981 4868 generic.go:334] "Generic (PLEG): container finished" podID="e3e9efc3-aa56-442a-9e27-44b72f11d3ad" containerID="8aad16845f92ec5322db4c77da6ca3cceb72bc33cbede6390eace86eb1d9829f" exitCode=137 Dec 01 17:45:59 crc kubenswrapper[4868]: I1201 17:45:59.907102 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e3e9efc3-aa56-442a-9e27-44b72f11d3ad","Type":"ContainerDied","Data":"8aad16845f92ec5322db4c77da6ca3cceb72bc33cbede6390eace86eb1d9829f"} Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.340137 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.499701 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-config-data\") pod \"e3e9efc3-aa56-442a-9e27-44b72f11d3ad\" (UID: \"e3e9efc3-aa56-442a-9e27-44b72f11d3ad\") " Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.499864 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-combined-ca-bundle\") pod \"e3e9efc3-aa56-442a-9e27-44b72f11d3ad\" (UID: \"e3e9efc3-aa56-442a-9e27-44b72f11d3ad\") " Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.499922 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfxz4\" (UniqueName: \"kubernetes.io/projected/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-kube-api-access-xfxz4\") pod \"e3e9efc3-aa56-442a-9e27-44b72f11d3ad\" (UID: \"e3e9efc3-aa56-442a-9e27-44b72f11d3ad\") " Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.505890 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-kube-api-access-xfxz4" (OuterVolumeSpecName: "kube-api-access-xfxz4") pod "e3e9efc3-aa56-442a-9e27-44b72f11d3ad" (UID: "e3e9efc3-aa56-442a-9e27-44b72f11d3ad"). InnerVolumeSpecName "kube-api-access-xfxz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.530691 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-config-data" (OuterVolumeSpecName: "config-data") pod "e3e9efc3-aa56-442a-9e27-44b72f11d3ad" (UID: "e3e9efc3-aa56-442a-9e27-44b72f11d3ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.531181 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3e9efc3-aa56-442a-9e27-44b72f11d3ad" (UID: "e3e9efc3-aa56-442a-9e27-44b72f11d3ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.602487 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.602525 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.602542 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfxz4\" (UniqueName: \"kubernetes.io/projected/e3e9efc3-aa56-442a-9e27-44b72f11d3ad-kube-api-access-xfxz4\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.916480 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e3e9efc3-aa56-442a-9e27-44b72f11d3ad","Type":"ContainerDied","Data":"bdbf7468d1357f01e99e6fc2c7a79da4a4a4eb28dccd45c9a40607321b71a44e"} Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.916538 4868 scope.go:117] "RemoveContainer" containerID="8aad16845f92ec5322db4c77da6ca3cceb72bc33cbede6390eace86eb1d9829f" Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.916587 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.957371 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.971600 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.988160 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 17:46:00 crc kubenswrapper[4868]: E1201 17:46:00.988668 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e9efc3-aa56-442a-9e27-44b72f11d3ad" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.988691 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e9efc3-aa56-442a-9e27-44b72f11d3ad" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.989030 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3e9efc3-aa56-442a-9e27-44b72f11d3ad" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.989812 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.993290 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.993532 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 01 17:46:00 crc kubenswrapper[4868]: I1201 17:46:00.994134 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.000818 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.119043 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4014a0ba-e992-4592-aafd-62e2c90e2ab0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4014a0ba-e992-4592-aafd-62e2c90e2ab0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.119129 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cjkn\" (UniqueName: \"kubernetes.io/projected/4014a0ba-e992-4592-aafd-62e2c90e2ab0-kube-api-access-4cjkn\") pod \"nova-cell1-novncproxy-0\" (UID: \"4014a0ba-e992-4592-aafd-62e2c90e2ab0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.119220 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4014a0ba-e992-4592-aafd-62e2c90e2ab0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4014a0ba-e992-4592-aafd-62e2c90e2ab0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.119249 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4014a0ba-e992-4592-aafd-62e2c90e2ab0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4014a0ba-e992-4592-aafd-62e2c90e2ab0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.119321 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4014a0ba-e992-4592-aafd-62e2c90e2ab0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4014a0ba-e992-4592-aafd-62e2c90e2ab0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.220805 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4014a0ba-e992-4592-aafd-62e2c90e2ab0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4014a0ba-e992-4592-aafd-62e2c90e2ab0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.220921 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cjkn\" (UniqueName: \"kubernetes.io/projected/4014a0ba-e992-4592-aafd-62e2c90e2ab0-kube-api-access-4cjkn\") pod \"nova-cell1-novncproxy-0\" (UID: \"4014a0ba-e992-4592-aafd-62e2c90e2ab0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.221047 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4014a0ba-e992-4592-aafd-62e2c90e2ab0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4014a0ba-e992-4592-aafd-62e2c90e2ab0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.221075 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4014a0ba-e992-4592-aafd-62e2c90e2ab0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4014a0ba-e992-4592-aafd-62e2c90e2ab0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.221144 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4014a0ba-e992-4592-aafd-62e2c90e2ab0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4014a0ba-e992-4592-aafd-62e2c90e2ab0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.224700 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4014a0ba-e992-4592-aafd-62e2c90e2ab0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4014a0ba-e992-4592-aafd-62e2c90e2ab0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.225187 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4014a0ba-e992-4592-aafd-62e2c90e2ab0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4014a0ba-e992-4592-aafd-62e2c90e2ab0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.226221 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4014a0ba-e992-4592-aafd-62e2c90e2ab0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4014a0ba-e992-4592-aafd-62e2c90e2ab0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.235733 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4014a0ba-e992-4592-aafd-62e2c90e2ab0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4014a0ba-e992-4592-aafd-62e2c90e2ab0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.237307 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cjkn\" (UniqueName: \"kubernetes.io/projected/4014a0ba-e992-4592-aafd-62e2c90e2ab0-kube-api-access-4cjkn\") pod \"nova-cell1-novncproxy-0\" (UID: \"4014a0ba-e992-4592-aafd-62e2c90e2ab0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.363717 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.787248 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 17:46:01 crc kubenswrapper[4868]: W1201 17:46:01.790531 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4014a0ba_e992_4592_aafd_62e2c90e2ab0.slice/crio-b7bccd0b8440c54f9f748fbe4591c52fbc1859e558993e7ebfe4f4b2d092fc55 WatchSource:0}: Error finding container b7bccd0b8440c54f9f748fbe4591c52fbc1859e558993e7ebfe4f4b2d092fc55: Status 404 returned error can't find the container with id b7bccd0b8440c54f9f748fbe4591c52fbc1859e558993e7ebfe4f4b2d092fc55 Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.933176 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4014a0ba-e992-4592-aafd-62e2c90e2ab0","Type":"ContainerStarted","Data":"b7bccd0b8440c54f9f748fbe4591c52fbc1859e558993e7ebfe4f4b2d092fc55"} Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.948050 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.948469 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.948498 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 17:46:01 crc kubenswrapper[4868]: I1201 17:46:01.951294 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 17:46:02 crc kubenswrapper[4868]: I1201 17:46:02.181844 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3e9efc3-aa56-442a-9e27-44b72f11d3ad" path="/var/lib/kubelet/pods/e3e9efc3-aa56-442a-9e27-44b72f11d3ad/volumes" Dec 01 17:46:02 crc kubenswrapper[4868]: I1201 17:46:02.943271 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4014a0ba-e992-4592-aafd-62e2c90e2ab0","Type":"ContainerStarted","Data":"53df50d44b21203d8ead2043a3f57a138db2874bc583e940ca52805912e994a7"} Dec 01 17:46:02 crc kubenswrapper[4868]: I1201 17:46:02.944513 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 17:46:02 crc kubenswrapper[4868]: I1201 17:46:02.948458 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 17:46:02 crc kubenswrapper[4868]: I1201 17:46:02.972875 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.97285104 podStartE2EDuration="2.97285104s" podCreationTimestamp="2025-12-01 17:46:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:46:02.964397155 +0000 UTC m=+1235.335507566" watchObservedRunningTime="2025-12-01 17:46:02.97285104 +0000 UTC m=+1235.343961451" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.137659 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-pkhvw"] Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.139398 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.166144 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-pkhvw"] Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.257370 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctkrj\" (UniqueName: \"kubernetes.io/projected/84e8b909-9eb9-4fdd-97fb-4b98256006aa-kube-api-access-ctkrj\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.257424 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.257800 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.258068 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-config\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.258287 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.258366 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.359562 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.359652 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-config\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.359712 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.359775 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.359803 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctkrj\" (UniqueName: \"kubernetes.io/projected/84e8b909-9eb9-4fdd-97fb-4b98256006aa-kube-api-access-ctkrj\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.359833 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.360878 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.361553 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.362187 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-config\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.362789 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.363357 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.398687 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctkrj\" (UniqueName: \"kubernetes.io/projected/84e8b909-9eb9-4fdd-97fb-4b98256006aa-kube-api-access-ctkrj\") pod \"dnsmasq-dns-cd5cbd7b9-pkhvw\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.474507 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:03 crc kubenswrapper[4868]: I1201 17:46:03.996897 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-pkhvw"] Dec 01 17:46:04 crc kubenswrapper[4868]: W1201 17:46:04.008572 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84e8b909_9eb9_4fdd_97fb_4b98256006aa.slice/crio-0ad25d1c081d4456e72ca96552eed495b1c2d19eca88b82b43531660572cb982 WatchSource:0}: Error finding container 0ad25d1c081d4456e72ca96552eed495b1c2d19eca88b82b43531660572cb982: Status 404 returned error can't find the container with id 0ad25d1c081d4456e72ca96552eed495b1c2d19eca88b82b43531660572cb982 Dec 01 17:46:04 crc kubenswrapper[4868]: I1201 17:46:04.942172 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:46:04 crc kubenswrapper[4868]: I1201 17:46:04.942774 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="ceilometer-central-agent" containerID="cri-o://5ab0302df9caf88ccebb6696b9836c30e56547f52571bfeb47b37170ede16727" gracePeriod=30 Dec 01 17:46:04 crc kubenswrapper[4868]: I1201 17:46:04.942890 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="ceilometer-notification-agent" containerID="cri-o://c633510a6fb594f0774bc484ba14c1ab5165955d6a447c972ce5854be1a6971f" gracePeriod=30 Dec 01 17:46:04 crc kubenswrapper[4868]: I1201 17:46:04.942897 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="proxy-httpd" containerID="cri-o://07f613806cd357e8e4387bd714d3cf2b9ec2c76343ac5d461d72829f4f1dc917" gracePeriod=30 Dec 01 17:46:04 crc kubenswrapper[4868]: I1201 17:46:04.942929 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="sg-core" containerID="cri-o://99a1b6fc34c2e67c253ca05527f7e655c2feb0a0b20a20e4ca6aef7ece34d702" gracePeriod=30 Dec 01 17:46:04 crc kubenswrapper[4868]: I1201 17:46:04.958433 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.197:3000/\": EOF" Dec 01 17:46:04 crc kubenswrapper[4868]: I1201 17:46:04.960819 4868 generic.go:334] "Generic (PLEG): container finished" podID="84e8b909-9eb9-4fdd-97fb-4b98256006aa" containerID="bd0ef08c1a95cec1b2f700e86f514dafb7ddd4f269f1b67a10250eac3896a6e4" exitCode=0 Dec 01 17:46:04 crc kubenswrapper[4868]: I1201 17:46:04.962145 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" event={"ID":"84e8b909-9eb9-4fdd-97fb-4b98256006aa","Type":"ContainerDied","Data":"bd0ef08c1a95cec1b2f700e86f514dafb7ddd4f269f1b67a10250eac3896a6e4"} Dec 01 17:46:04 crc kubenswrapper[4868]: I1201 17:46:04.962178 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" event={"ID":"84e8b909-9eb9-4fdd-97fb-4b98256006aa","Type":"ContainerStarted","Data":"0ad25d1c081d4456e72ca96552eed495b1c2d19eca88b82b43531660572cb982"} Dec 01 17:46:05 crc kubenswrapper[4868]: I1201 17:46:05.916451 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:46:05 crc kubenswrapper[4868]: I1201 17:46:05.972006 4868 generic.go:334] "Generic (PLEG): container finished" podID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerID="07f613806cd357e8e4387bd714d3cf2b9ec2c76343ac5d461d72829f4f1dc917" exitCode=0 Dec 01 17:46:05 crc kubenswrapper[4868]: I1201 17:46:05.972044 4868 generic.go:334] "Generic (PLEG): container finished" podID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerID="99a1b6fc34c2e67c253ca05527f7e655c2feb0a0b20a20e4ca6aef7ece34d702" exitCode=2 Dec 01 17:46:05 crc kubenswrapper[4868]: I1201 17:46:05.972054 4868 generic.go:334] "Generic (PLEG): container finished" podID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerID="5ab0302df9caf88ccebb6696b9836c30e56547f52571bfeb47b37170ede16727" exitCode=0 Dec 01 17:46:05 crc kubenswrapper[4868]: I1201 17:46:05.972084 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c523a5-0e2d-4e46-b2c7-516ab7cf2617","Type":"ContainerDied","Data":"07f613806cd357e8e4387bd714d3cf2b9ec2c76343ac5d461d72829f4f1dc917"} Dec 01 17:46:05 crc kubenswrapper[4868]: I1201 17:46:05.972128 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c523a5-0e2d-4e46-b2c7-516ab7cf2617","Type":"ContainerDied","Data":"99a1b6fc34c2e67c253ca05527f7e655c2feb0a0b20a20e4ca6aef7ece34d702"} Dec 01 17:46:05 crc kubenswrapper[4868]: I1201 17:46:05.972138 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c523a5-0e2d-4e46-b2c7-516ab7cf2617","Type":"ContainerDied","Data":"5ab0302df9caf88ccebb6696b9836c30e56547f52571bfeb47b37170ede16727"} Dec 01 17:46:05 crc kubenswrapper[4868]: I1201 17:46:05.973984 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" event={"ID":"84e8b909-9eb9-4fdd-97fb-4b98256006aa","Type":"ContainerStarted","Data":"093bbc91a298e5f48e40e82040e6a5b1956f89305df649897d67c8242fa90136"} Dec 01 17:46:05 crc kubenswrapper[4868]: I1201 17:46:05.974123 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="430e5d96-e42d-433f-b311-39a7ab09a98c" containerName="nova-api-log" containerID="cri-o://433ab3fd5801bc8a68904a86200c42f15979352c6babf8f6fc1cf60ebf1fc7bf" gracePeriod=30 Dec 01 17:46:05 crc kubenswrapper[4868]: I1201 17:46:05.974210 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="430e5d96-e42d-433f-b311-39a7ab09a98c" containerName="nova-api-api" containerID="cri-o://0105f3c449df8f1a4888fdb10c28c6358b67a6322627d5ad73a2e7d1f3b73344" gracePeriod=30 Dec 01 17:46:06 crc kubenswrapper[4868]: I1201 17:46:06.016811 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" podStartSLOduration=3.016795075 podStartE2EDuration="3.016795075s" podCreationTimestamp="2025-12-01 17:46:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:46:06.001754303 +0000 UTC m=+1238.372864714" watchObservedRunningTime="2025-12-01 17:46:06.016795075 +0000 UTC m=+1238.387905486" Dec 01 17:46:06 crc kubenswrapper[4868]: I1201 17:46:06.363781 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:06 crc kubenswrapper[4868]: I1201 17:46:06.983109 4868 generic.go:334] "Generic (PLEG): container finished" podID="430e5d96-e42d-433f-b311-39a7ab09a98c" containerID="433ab3fd5801bc8a68904a86200c42f15979352c6babf8f6fc1cf60ebf1fc7bf" exitCode=143 Dec 01 17:46:06 crc kubenswrapper[4868]: I1201 17:46:06.983189 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"430e5d96-e42d-433f-b311-39a7ab09a98c","Type":"ContainerDied","Data":"433ab3fd5801bc8a68904a86200c42f15979352c6babf8f6fc1cf60ebf1fc7bf"} Dec 01 17:46:06 crc kubenswrapper[4868]: I1201 17:46:06.983960 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.561667 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.667110 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-scripts\") pod \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.667180 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-combined-ca-bundle\") pod \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.667372 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-run-httpd\") pod \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.667456 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f57gs\" (UniqueName: \"kubernetes.io/projected/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-kube-api-access-f57gs\") pod \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.667483 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-config-data\") pod \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.667532 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-log-httpd\") pod \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.667554 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-ceilometer-tls-certs\") pod \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.667578 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-sg-core-conf-yaml\") pod \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\" (UID: \"11c523a5-0e2d-4e46-b2c7-516ab7cf2617\") " Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.667755 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "11c523a5-0e2d-4e46-b2c7-516ab7cf2617" (UID: "11c523a5-0e2d-4e46-b2c7-516ab7cf2617"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.667882 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "11c523a5-0e2d-4e46-b2c7-516ab7cf2617" (UID: "11c523a5-0e2d-4e46-b2c7-516ab7cf2617"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.668441 4868 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.668459 4868 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.673635 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-scripts" (OuterVolumeSpecName: "scripts") pod "11c523a5-0e2d-4e46-b2c7-516ab7cf2617" (UID: "11c523a5-0e2d-4e46-b2c7-516ab7cf2617"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.673996 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-kube-api-access-f57gs" (OuterVolumeSpecName: "kube-api-access-f57gs") pod "11c523a5-0e2d-4e46-b2c7-516ab7cf2617" (UID: "11c523a5-0e2d-4e46-b2c7-516ab7cf2617"). InnerVolumeSpecName "kube-api-access-f57gs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.719491 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "11c523a5-0e2d-4e46-b2c7-516ab7cf2617" (UID: "11c523a5-0e2d-4e46-b2c7-516ab7cf2617"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.723685 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "11c523a5-0e2d-4e46-b2c7-516ab7cf2617" (UID: "11c523a5-0e2d-4e46-b2c7-516ab7cf2617"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.757276 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11c523a5-0e2d-4e46-b2c7-516ab7cf2617" (UID: "11c523a5-0e2d-4e46-b2c7-516ab7cf2617"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.773224 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f57gs\" (UniqueName: \"kubernetes.io/projected/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-kube-api-access-f57gs\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.773263 4868 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.773276 4868 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.773290 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.773302 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.790353 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-config-data" (OuterVolumeSpecName: "config-data") pod "11c523a5-0e2d-4e46-b2c7-516ab7cf2617" (UID: "11c523a5-0e2d-4e46-b2c7-516ab7cf2617"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:07 crc kubenswrapper[4868]: I1201 17:46:07.875447 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c523a5-0e2d-4e46-b2c7-516ab7cf2617-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.005698 4868 generic.go:334] "Generic (PLEG): container finished" podID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerID="c633510a6fb594f0774bc484ba14c1ab5165955d6a447c972ce5854be1a6971f" exitCode=0 Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.005757 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c523a5-0e2d-4e46-b2c7-516ab7cf2617","Type":"ContainerDied","Data":"c633510a6fb594f0774bc484ba14c1ab5165955d6a447c972ce5854be1a6971f"} Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.005800 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c523a5-0e2d-4e46-b2c7-516ab7cf2617","Type":"ContainerDied","Data":"0d9599ab48901d40f5f0c3a01a8ca30e9236b322f60a5febf7ccc2377eb96899"} Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.005823 4868 scope.go:117] "RemoveContainer" containerID="07f613806cd357e8e4387bd714d3cf2b9ec2c76343ac5d461d72829f4f1dc917" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.005822 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.026385 4868 scope.go:117] "RemoveContainer" containerID="99a1b6fc34c2e67c253ca05527f7e655c2feb0a0b20a20e4ca6aef7ece34d702" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.048826 4868 scope.go:117] "RemoveContainer" containerID="c633510a6fb594f0774bc484ba14c1ab5165955d6a447c972ce5854be1a6971f" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.049663 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.066849 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.079862 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:46:08 crc kubenswrapper[4868]: E1201 17:46:08.080342 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="ceilometer-notification-agent" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.080368 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="ceilometer-notification-agent" Dec 01 17:46:08 crc kubenswrapper[4868]: E1201 17:46:08.080392 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="proxy-httpd" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.080402 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="proxy-httpd" Dec 01 17:46:08 crc kubenswrapper[4868]: E1201 17:46:08.080424 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="sg-core" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.080429 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="sg-core" Dec 01 17:46:08 crc kubenswrapper[4868]: E1201 17:46:08.080440 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="ceilometer-central-agent" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.080446 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="ceilometer-central-agent" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.080669 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="ceilometer-central-agent" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.080685 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="sg-core" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.080701 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="proxy-httpd" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.080729 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" containerName="ceilometer-notification-agent" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.082851 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.085314 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.085429 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.086894 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.088192 4868 scope.go:117] "RemoveContainer" containerID="5ab0302df9caf88ccebb6696b9836c30e56547f52571bfeb47b37170ede16727" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.096912 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.133201 4868 scope.go:117] "RemoveContainer" containerID="07f613806cd357e8e4387bd714d3cf2b9ec2c76343ac5d461d72829f4f1dc917" Dec 01 17:46:08 crc kubenswrapper[4868]: E1201 17:46:08.133708 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07f613806cd357e8e4387bd714d3cf2b9ec2c76343ac5d461d72829f4f1dc917\": container with ID starting with 07f613806cd357e8e4387bd714d3cf2b9ec2c76343ac5d461d72829f4f1dc917 not found: ID does not exist" containerID="07f613806cd357e8e4387bd714d3cf2b9ec2c76343ac5d461d72829f4f1dc917" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.133741 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07f613806cd357e8e4387bd714d3cf2b9ec2c76343ac5d461d72829f4f1dc917"} err="failed to get container status \"07f613806cd357e8e4387bd714d3cf2b9ec2c76343ac5d461d72829f4f1dc917\": rpc error: code = NotFound desc = could not find container \"07f613806cd357e8e4387bd714d3cf2b9ec2c76343ac5d461d72829f4f1dc917\": container with ID starting with 07f613806cd357e8e4387bd714d3cf2b9ec2c76343ac5d461d72829f4f1dc917 not found: ID does not exist" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.133761 4868 scope.go:117] "RemoveContainer" containerID="99a1b6fc34c2e67c253ca05527f7e655c2feb0a0b20a20e4ca6aef7ece34d702" Dec 01 17:46:08 crc kubenswrapper[4868]: E1201 17:46:08.134360 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99a1b6fc34c2e67c253ca05527f7e655c2feb0a0b20a20e4ca6aef7ece34d702\": container with ID starting with 99a1b6fc34c2e67c253ca05527f7e655c2feb0a0b20a20e4ca6aef7ece34d702 not found: ID does not exist" containerID="99a1b6fc34c2e67c253ca05527f7e655c2feb0a0b20a20e4ca6aef7ece34d702" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.134383 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99a1b6fc34c2e67c253ca05527f7e655c2feb0a0b20a20e4ca6aef7ece34d702"} err="failed to get container status \"99a1b6fc34c2e67c253ca05527f7e655c2feb0a0b20a20e4ca6aef7ece34d702\": rpc error: code = NotFound desc = could not find container \"99a1b6fc34c2e67c253ca05527f7e655c2feb0a0b20a20e4ca6aef7ece34d702\": container with ID starting with 99a1b6fc34c2e67c253ca05527f7e655c2feb0a0b20a20e4ca6aef7ece34d702 not found: ID does not exist" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.134395 4868 scope.go:117] "RemoveContainer" containerID="c633510a6fb594f0774bc484ba14c1ab5165955d6a447c972ce5854be1a6971f" Dec 01 17:46:08 crc kubenswrapper[4868]: E1201 17:46:08.135885 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c633510a6fb594f0774bc484ba14c1ab5165955d6a447c972ce5854be1a6971f\": container with ID starting with c633510a6fb594f0774bc484ba14c1ab5165955d6a447c972ce5854be1a6971f not found: ID does not exist" containerID="c633510a6fb594f0774bc484ba14c1ab5165955d6a447c972ce5854be1a6971f" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.135906 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c633510a6fb594f0774bc484ba14c1ab5165955d6a447c972ce5854be1a6971f"} err="failed to get container status \"c633510a6fb594f0774bc484ba14c1ab5165955d6a447c972ce5854be1a6971f\": rpc error: code = NotFound desc = could not find container \"c633510a6fb594f0774bc484ba14c1ab5165955d6a447c972ce5854be1a6971f\": container with ID starting with c633510a6fb594f0774bc484ba14c1ab5165955d6a447c972ce5854be1a6971f not found: ID does not exist" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.135983 4868 scope.go:117] "RemoveContainer" containerID="5ab0302df9caf88ccebb6696b9836c30e56547f52571bfeb47b37170ede16727" Dec 01 17:46:08 crc kubenswrapper[4868]: E1201 17:46:08.136454 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ab0302df9caf88ccebb6696b9836c30e56547f52571bfeb47b37170ede16727\": container with ID starting with 5ab0302df9caf88ccebb6696b9836c30e56547f52571bfeb47b37170ede16727 not found: ID does not exist" containerID="5ab0302df9caf88ccebb6696b9836c30e56547f52571bfeb47b37170ede16727" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.136478 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ab0302df9caf88ccebb6696b9836c30e56547f52571bfeb47b37170ede16727"} err="failed to get container status \"5ab0302df9caf88ccebb6696b9836c30e56547f52571bfeb47b37170ede16727\": rpc error: code = NotFound desc = could not find container \"5ab0302df9caf88ccebb6696b9836c30e56547f52571bfeb47b37170ede16727\": container with ID starting with 5ab0302df9caf88ccebb6696b9836c30e56547f52571bfeb47b37170ede16727 not found: ID does not exist" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.179772 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0831ed8f-c5a0-4273-a32d-4f338a325073-log-httpd\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.179922 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j8n4\" (UniqueName: \"kubernetes.io/projected/0831ed8f-c5a0-4273-a32d-4f338a325073-kube-api-access-6j8n4\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.180474 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0831ed8f-c5a0-4273-a32d-4f338a325073-run-httpd\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.180631 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0831ed8f-c5a0-4273-a32d-4f338a325073-config-data\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.180654 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0831ed8f-c5a0-4273-a32d-4f338a325073-scripts\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.180737 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0831ed8f-c5a0-4273-a32d-4f338a325073-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.180808 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0831ed8f-c5a0-4273-a32d-4f338a325073-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.180846 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0831ed8f-c5a0-4273-a32d-4f338a325073-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.191784 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11c523a5-0e2d-4e46-b2c7-516ab7cf2617" path="/var/lib/kubelet/pods/11c523a5-0e2d-4e46-b2c7-516ab7cf2617/volumes" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.282422 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0831ed8f-c5a0-4273-a32d-4f338a325073-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.282492 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0831ed8f-c5a0-4273-a32d-4f338a325073-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.282536 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0831ed8f-c5a0-4273-a32d-4f338a325073-log-httpd\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.282639 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j8n4\" (UniqueName: \"kubernetes.io/projected/0831ed8f-c5a0-4273-a32d-4f338a325073-kube-api-access-6j8n4\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.282662 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0831ed8f-c5a0-4273-a32d-4f338a325073-run-httpd\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.282695 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0831ed8f-c5a0-4273-a32d-4f338a325073-config-data\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.282709 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0831ed8f-c5a0-4273-a32d-4f338a325073-scripts\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.282747 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0831ed8f-c5a0-4273-a32d-4f338a325073-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.284436 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0831ed8f-c5a0-4273-a32d-4f338a325073-run-httpd\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.284754 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0831ed8f-c5a0-4273-a32d-4f338a325073-log-httpd\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.288451 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0831ed8f-c5a0-4273-a32d-4f338a325073-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.289092 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0831ed8f-c5a0-4273-a32d-4f338a325073-scripts\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.293015 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0831ed8f-c5a0-4273-a32d-4f338a325073-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.296404 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0831ed8f-c5a0-4273-a32d-4f338a325073-config-data\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.301884 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j8n4\" (UniqueName: \"kubernetes.io/projected/0831ed8f-c5a0-4273-a32d-4f338a325073-kube-api-access-6j8n4\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.304597 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0831ed8f-c5a0-4273-a32d-4f338a325073-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0831ed8f-c5a0-4273-a32d-4f338a325073\") " pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.418710 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 17:46:08 crc kubenswrapper[4868]: W1201 17:46:08.902139 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0831ed8f_c5a0_4273_a32d_4f338a325073.slice/crio-9afa9cdad843dcf5e312026bc4c11aec7960d06ec85ae58630793dfb60416e24 WatchSource:0}: Error finding container 9afa9cdad843dcf5e312026bc4c11aec7960d06ec85ae58630793dfb60416e24: Status 404 returned error can't find the container with id 9afa9cdad843dcf5e312026bc4c11aec7960d06ec85ae58630793dfb60416e24 Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.904786 4868 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 17:46:08 crc kubenswrapper[4868]: I1201 17:46:08.910675 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 17:46:09 crc kubenswrapper[4868]: I1201 17:46:09.016390 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0831ed8f-c5a0-4273-a32d-4f338a325073","Type":"ContainerStarted","Data":"9afa9cdad843dcf5e312026bc4c11aec7960d06ec85ae58630793dfb60416e24"} Dec 01 17:46:09 crc kubenswrapper[4868]: I1201 17:46:09.591083 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:46:09 crc kubenswrapper[4868]: I1201 17:46:09.710108 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430e5d96-e42d-433f-b311-39a7ab09a98c-combined-ca-bundle\") pod \"430e5d96-e42d-433f-b311-39a7ab09a98c\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " Dec 01 17:46:09 crc kubenswrapper[4868]: I1201 17:46:09.710686 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvp7s\" (UniqueName: \"kubernetes.io/projected/430e5d96-e42d-433f-b311-39a7ab09a98c-kube-api-access-nvp7s\") pod \"430e5d96-e42d-433f-b311-39a7ab09a98c\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " Dec 01 17:46:09 crc kubenswrapper[4868]: I1201 17:46:09.710802 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430e5d96-e42d-433f-b311-39a7ab09a98c-logs\") pod \"430e5d96-e42d-433f-b311-39a7ab09a98c\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " Dec 01 17:46:09 crc kubenswrapper[4868]: I1201 17:46:09.710961 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430e5d96-e42d-433f-b311-39a7ab09a98c-config-data\") pod \"430e5d96-e42d-433f-b311-39a7ab09a98c\" (UID: \"430e5d96-e42d-433f-b311-39a7ab09a98c\") " Dec 01 17:46:09 crc kubenswrapper[4868]: I1201 17:46:09.713688 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/430e5d96-e42d-433f-b311-39a7ab09a98c-logs" (OuterVolumeSpecName: "logs") pod "430e5d96-e42d-433f-b311-39a7ab09a98c" (UID: "430e5d96-e42d-433f-b311-39a7ab09a98c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:46:09 crc kubenswrapper[4868]: I1201 17:46:09.718377 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/430e5d96-e42d-433f-b311-39a7ab09a98c-kube-api-access-nvp7s" (OuterVolumeSpecName: "kube-api-access-nvp7s") pod "430e5d96-e42d-433f-b311-39a7ab09a98c" (UID: "430e5d96-e42d-433f-b311-39a7ab09a98c"). InnerVolumeSpecName "kube-api-access-nvp7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:46:09 crc kubenswrapper[4868]: I1201 17:46:09.760050 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/430e5d96-e42d-433f-b311-39a7ab09a98c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "430e5d96-e42d-433f-b311-39a7ab09a98c" (UID: "430e5d96-e42d-433f-b311-39a7ab09a98c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:09 crc kubenswrapper[4868]: I1201 17:46:09.762630 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/430e5d96-e42d-433f-b311-39a7ab09a98c-config-data" (OuterVolumeSpecName: "config-data") pod "430e5d96-e42d-433f-b311-39a7ab09a98c" (UID: "430e5d96-e42d-433f-b311-39a7ab09a98c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:09 crc kubenswrapper[4868]: I1201 17:46:09.812663 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430e5d96-e42d-433f-b311-39a7ab09a98c-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:09 crc kubenswrapper[4868]: I1201 17:46:09.812779 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430e5d96-e42d-433f-b311-39a7ab09a98c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:09 crc kubenswrapper[4868]: I1201 17:46:09.812794 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvp7s\" (UniqueName: \"kubernetes.io/projected/430e5d96-e42d-433f-b311-39a7ab09a98c-kube-api-access-nvp7s\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:09 crc kubenswrapper[4868]: I1201 17:46:09.812823 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430e5d96-e42d-433f-b311-39a7ab09a98c-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.030261 4868 generic.go:334] "Generic (PLEG): container finished" podID="430e5d96-e42d-433f-b311-39a7ab09a98c" containerID="0105f3c449df8f1a4888fdb10c28c6358b67a6322627d5ad73a2e7d1f3b73344" exitCode=0 Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.030317 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.030350 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"430e5d96-e42d-433f-b311-39a7ab09a98c","Type":"ContainerDied","Data":"0105f3c449df8f1a4888fdb10c28c6358b67a6322627d5ad73a2e7d1f3b73344"} Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.030410 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"430e5d96-e42d-433f-b311-39a7ab09a98c","Type":"ContainerDied","Data":"5e1f593e47118489217ee5e49cc3a196e5725bc8a40a66ef069a19e6f99ea093"} Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.030438 4868 scope.go:117] "RemoveContainer" containerID="0105f3c449df8f1a4888fdb10c28c6358b67a6322627d5ad73a2e7d1f3b73344" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.033026 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0831ed8f-c5a0-4273-a32d-4f338a325073","Type":"ContainerStarted","Data":"f6414f883be2f8e032deabb3fe0a62fc743db75cdba1b799db33362150f452b9"} Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.074059 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.083098 4868 scope.go:117] "RemoveContainer" containerID="433ab3fd5801bc8a68904a86200c42f15979352c6babf8f6fc1cf60ebf1fc7bf" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.088065 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.108433 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 17:46:10 crc kubenswrapper[4868]: E1201 17:46:10.109105 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="430e5d96-e42d-433f-b311-39a7ab09a98c" containerName="nova-api-api" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.109135 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="430e5d96-e42d-433f-b311-39a7ab09a98c" containerName="nova-api-api" Dec 01 17:46:10 crc kubenswrapper[4868]: E1201 17:46:10.109174 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="430e5d96-e42d-433f-b311-39a7ab09a98c" containerName="nova-api-log" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.109186 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="430e5d96-e42d-433f-b311-39a7ab09a98c" containerName="nova-api-log" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.109764 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="430e5d96-e42d-433f-b311-39a7ab09a98c" containerName="nova-api-log" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.109854 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="430e5d96-e42d-433f-b311-39a7ab09a98c" containerName="nova-api-api" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.114219 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.119065 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.119357 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.119687 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.125200 4868 scope.go:117] "RemoveContainer" containerID="0105f3c449df8f1a4888fdb10c28c6358b67a6322627d5ad73a2e7d1f3b73344" Dec 01 17:46:10 crc kubenswrapper[4868]: E1201 17:46:10.126450 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0105f3c449df8f1a4888fdb10c28c6358b67a6322627d5ad73a2e7d1f3b73344\": container with ID starting with 0105f3c449df8f1a4888fdb10c28c6358b67a6322627d5ad73a2e7d1f3b73344 not found: ID does not exist" containerID="0105f3c449df8f1a4888fdb10c28c6358b67a6322627d5ad73a2e7d1f3b73344" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.126507 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0105f3c449df8f1a4888fdb10c28c6358b67a6322627d5ad73a2e7d1f3b73344"} err="failed to get container status \"0105f3c449df8f1a4888fdb10c28c6358b67a6322627d5ad73a2e7d1f3b73344\": rpc error: code = NotFound desc = could not find container \"0105f3c449df8f1a4888fdb10c28c6358b67a6322627d5ad73a2e7d1f3b73344\": container with ID starting with 0105f3c449df8f1a4888fdb10c28c6358b67a6322627d5ad73a2e7d1f3b73344 not found: ID does not exist" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.126531 4868 scope.go:117] "RemoveContainer" containerID="433ab3fd5801bc8a68904a86200c42f15979352c6babf8f6fc1cf60ebf1fc7bf" Dec 01 17:46:10 crc kubenswrapper[4868]: E1201 17:46:10.128763 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"433ab3fd5801bc8a68904a86200c42f15979352c6babf8f6fc1cf60ebf1fc7bf\": container with ID starting with 433ab3fd5801bc8a68904a86200c42f15979352c6babf8f6fc1cf60ebf1fc7bf not found: ID does not exist" containerID="433ab3fd5801bc8a68904a86200c42f15979352c6babf8f6fc1cf60ebf1fc7bf" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.128823 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"433ab3fd5801bc8a68904a86200c42f15979352c6babf8f6fc1cf60ebf1fc7bf"} err="failed to get container status \"433ab3fd5801bc8a68904a86200c42f15979352c6babf8f6fc1cf60ebf1fc7bf\": rpc error: code = NotFound desc = could not find container \"433ab3fd5801bc8a68904a86200c42f15979352c6babf8f6fc1cf60ebf1fc7bf\": container with ID starting with 433ab3fd5801bc8a68904a86200c42f15979352c6babf8f6fc1cf60ebf1fc7bf not found: ID does not exist" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.146743 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.193987 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="430e5d96-e42d-433f-b311-39a7ab09a98c" path="/var/lib/kubelet/pods/430e5d96-e42d-433f-b311-39a7ab09a98c/volumes" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.221720 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-config-data\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.221809 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.221837 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.221890 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bw52\" (UniqueName: \"kubernetes.io/projected/ccced3dc-1af6-4959-a40f-c4f0dd654fca-kube-api-access-7bw52\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.221987 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccced3dc-1af6-4959-a40f-c4f0dd654fca-logs\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.222062 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-public-tls-certs\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.324013 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-config-data\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.324078 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.324101 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.324125 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bw52\" (UniqueName: \"kubernetes.io/projected/ccced3dc-1af6-4959-a40f-c4f0dd654fca-kube-api-access-7bw52\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.324179 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccced3dc-1af6-4959-a40f-c4f0dd654fca-logs\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.324235 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-public-tls-certs\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.325551 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccced3dc-1af6-4959-a40f-c4f0dd654fca-logs\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.330170 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.330205 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-public-tls-certs\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.338538 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-config-data\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.340310 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.345631 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bw52\" (UniqueName: \"kubernetes.io/projected/ccced3dc-1af6-4959-a40f-c4f0dd654fca-kube-api-access-7bw52\") pod \"nova-api-0\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.437856 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:46:10 crc kubenswrapper[4868]: I1201 17:46:10.905756 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:46:10 crc kubenswrapper[4868]: W1201 17:46:10.909757 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccced3dc_1af6_4959_a40f_c4f0dd654fca.slice/crio-53b9168409f25ad2a933f73a3ee238813660706197252cda8925fc89ef5e630b WatchSource:0}: Error finding container 53b9168409f25ad2a933f73a3ee238813660706197252cda8925fc89ef5e630b: Status 404 returned error can't find the container with id 53b9168409f25ad2a933f73a3ee238813660706197252cda8925fc89ef5e630b Dec 01 17:46:11 crc kubenswrapper[4868]: I1201 17:46:11.046008 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0831ed8f-c5a0-4273-a32d-4f338a325073","Type":"ContainerStarted","Data":"e4abe6d30f982c2dff157af2321b79706a61eecb263dad32c86e1f77dda9713b"} Dec 01 17:46:11 crc kubenswrapper[4868]: I1201 17:46:11.047754 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ccced3dc-1af6-4959-a40f-c4f0dd654fca","Type":"ContainerStarted","Data":"53b9168409f25ad2a933f73a3ee238813660706197252cda8925fc89ef5e630b"} Dec 01 17:46:11 crc kubenswrapper[4868]: I1201 17:46:11.363975 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:11 crc kubenswrapper[4868]: I1201 17:46:11.391763 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.061112 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0831ed8f-c5a0-4273-a32d-4f338a325073","Type":"ContainerStarted","Data":"db2f0691d2041b658e6f9230162aee994b1ba7faa95ceba46a0f907f89044eb3"} Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.064879 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ccced3dc-1af6-4959-a40f-c4f0dd654fca","Type":"ContainerStarted","Data":"b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11"} Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.064918 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ccced3dc-1af6-4959-a40f-c4f0dd654fca","Type":"ContainerStarted","Data":"0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b"} Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.084366 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.100850 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.100826767 podStartE2EDuration="2.100826767s" podCreationTimestamp="2025-12-01 17:46:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:46:12.08638464 +0000 UTC m=+1244.457495051" watchObservedRunningTime="2025-12-01 17:46:12.100826767 +0000 UTC m=+1244.471937178" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.242172 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-pjj7m"] Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.243519 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.246765 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.252713 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.261038 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-pjj7m"] Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.361619 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-config-data\") pod \"nova-cell1-cell-mapping-pjj7m\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.361664 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pjj7m\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.361692 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfr24\" (UniqueName: \"kubernetes.io/projected/491393ef-ffbb-4a81-a477-5856765b7a48-kube-api-access-dfr24\") pod \"nova-cell1-cell-mapping-pjj7m\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.361713 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-scripts\") pod \"nova-cell1-cell-mapping-pjj7m\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.463860 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-config-data\") pod \"nova-cell1-cell-mapping-pjj7m\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.464265 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pjj7m\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.464358 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfr24\" (UniqueName: \"kubernetes.io/projected/491393ef-ffbb-4a81-a477-5856765b7a48-kube-api-access-dfr24\") pod \"nova-cell1-cell-mapping-pjj7m\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.464422 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-scripts\") pod \"nova-cell1-cell-mapping-pjj7m\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.468704 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pjj7m\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.469168 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-scripts\") pod \"nova-cell1-cell-mapping-pjj7m\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.471282 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-config-data\") pod \"nova-cell1-cell-mapping-pjj7m\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.485059 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfr24\" (UniqueName: \"kubernetes.io/projected/491393ef-ffbb-4a81-a477-5856765b7a48-kube-api-access-dfr24\") pod \"nova-cell1-cell-mapping-pjj7m\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:12 crc kubenswrapper[4868]: I1201 17:46:12.559674 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:13 crc kubenswrapper[4868]: W1201 17:46:13.078216 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod491393ef_ffbb_4a81_a477_5856765b7a48.slice/crio-110bad1c3880de1db14729d69a4fa7590d3266de7d6265dbb41b991840d635f6 WatchSource:0}: Error finding container 110bad1c3880de1db14729d69a4fa7590d3266de7d6265dbb41b991840d635f6: Status 404 returned error can't find the container with id 110bad1c3880de1db14729d69a4fa7590d3266de7d6265dbb41b991840d635f6 Dec 01 17:46:13 crc kubenswrapper[4868]: I1201 17:46:13.095760 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-pjj7m"] Dec 01 17:46:13 crc kubenswrapper[4868]: I1201 17:46:13.475908 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:46:13 crc kubenswrapper[4868]: I1201 17:46:13.561590 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-z2bdk"] Dec 01 17:46:13 crc kubenswrapper[4868]: I1201 17:46:13.561799 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" podUID="19e9c57b-8750-414f-b95b-3938c1356ed0" containerName="dnsmasq-dns" containerID="cri-o://68c50850b354708778d1c5f0bf21257e977a75918b3a2edeb8bcb5d0ae9e3f23" gracePeriod=10 Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.096314 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0831ed8f-c5a0-4273-a32d-4f338a325073","Type":"ContainerStarted","Data":"b4838fab240173170a8f331afa56a1fae8f5bd8afb9181fed0319b6f64773f08"} Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.096622 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.105233 4868 generic.go:334] "Generic (PLEG): container finished" podID="19e9c57b-8750-414f-b95b-3938c1356ed0" containerID="68c50850b354708778d1c5f0bf21257e977a75918b3a2edeb8bcb5d0ae9e3f23" exitCode=0 Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.105324 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" event={"ID":"19e9c57b-8750-414f-b95b-3938c1356ed0","Type":"ContainerDied","Data":"68c50850b354708778d1c5f0bf21257e977a75918b3a2edeb8bcb5d0ae9e3f23"} Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.105350 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" event={"ID":"19e9c57b-8750-414f-b95b-3938c1356ed0","Type":"ContainerDied","Data":"d410126333bdad254e9d4c334cef6d36bbac6eed618d2dc6a58eac27f2145a11"} Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.105361 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d410126333bdad254e9d4c334cef6d36bbac6eed618d2dc6a58eac27f2145a11" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.107305 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pjj7m" event={"ID":"491393ef-ffbb-4a81-a477-5856765b7a48","Type":"ContainerStarted","Data":"98997845d571b01016aa693a291d1d21d2117a71f509381ce801cc152937438a"} Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.107331 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pjj7m" event={"ID":"491393ef-ffbb-4a81-a477-5856765b7a48","Type":"ContainerStarted","Data":"110bad1c3880de1db14729d69a4fa7590d3266de7d6265dbb41b991840d635f6"} Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.116700 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.983882897 podStartE2EDuration="6.116682422s" podCreationTimestamp="2025-12-01 17:46:08 +0000 UTC" firstStartedPulling="2025-12-01 17:46:08.90457535 +0000 UTC m=+1241.275685761" lastFinishedPulling="2025-12-01 17:46:13.037374875 +0000 UTC m=+1245.408485286" observedRunningTime="2025-12-01 17:46:14.116165099 +0000 UTC m=+1246.487275530" watchObservedRunningTime="2025-12-01 17:46:14.116682422 +0000 UTC m=+1246.487792833" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.123684 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.141757 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-pjj7m" podStartSLOduration=2.141736328 podStartE2EDuration="2.141736328s" podCreationTimestamp="2025-12-01 17:46:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:46:14.137423829 +0000 UTC m=+1246.508534240" watchObservedRunningTime="2025-12-01 17:46:14.141736328 +0000 UTC m=+1246.512846739" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.199641 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-dns-svc\") pod \"19e9c57b-8750-414f-b95b-3938c1356ed0\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.199761 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-ovsdbserver-nb\") pod \"19e9c57b-8750-414f-b95b-3938c1356ed0\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.200007 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-config\") pod \"19e9c57b-8750-414f-b95b-3938c1356ed0\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.200072 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-dns-swift-storage-0\") pod \"19e9c57b-8750-414f-b95b-3938c1356ed0\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.200152 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-ovsdbserver-sb\") pod \"19e9c57b-8750-414f-b95b-3938c1356ed0\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.200225 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c9dl\" (UniqueName: \"kubernetes.io/projected/19e9c57b-8750-414f-b95b-3938c1356ed0-kube-api-access-5c9dl\") pod \"19e9c57b-8750-414f-b95b-3938c1356ed0\" (UID: \"19e9c57b-8750-414f-b95b-3938c1356ed0\") " Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.245498 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19e9c57b-8750-414f-b95b-3938c1356ed0-kube-api-access-5c9dl" (OuterVolumeSpecName: "kube-api-access-5c9dl") pod "19e9c57b-8750-414f-b95b-3938c1356ed0" (UID: "19e9c57b-8750-414f-b95b-3938c1356ed0"). InnerVolumeSpecName "kube-api-access-5c9dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.275443 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "19e9c57b-8750-414f-b95b-3938c1356ed0" (UID: "19e9c57b-8750-414f-b95b-3938c1356ed0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.283231 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "19e9c57b-8750-414f-b95b-3938c1356ed0" (UID: "19e9c57b-8750-414f-b95b-3938c1356ed0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.293705 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "19e9c57b-8750-414f-b95b-3938c1356ed0" (UID: "19e9c57b-8750-414f-b95b-3938c1356ed0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.305337 4868 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.305365 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c9dl\" (UniqueName: \"kubernetes.io/projected/19e9c57b-8750-414f-b95b-3938c1356ed0-kube-api-access-5c9dl\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.305376 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.305385 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.307629 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-config" (OuterVolumeSpecName: "config") pod "19e9c57b-8750-414f-b95b-3938c1356ed0" (UID: "19e9c57b-8750-414f-b95b-3938c1356ed0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.309761 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "19e9c57b-8750-414f-b95b-3938c1356ed0" (UID: "19e9c57b-8750-414f-b95b-3938c1356ed0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.406674 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:14 crc kubenswrapper[4868]: I1201 17:46:14.406715 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19e9c57b-8750-414f-b95b-3938c1356ed0-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:15 crc kubenswrapper[4868]: I1201 17:46:15.114935 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" Dec 01 17:46:15 crc kubenswrapper[4868]: I1201 17:46:15.155665 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-z2bdk"] Dec 01 17:46:15 crc kubenswrapper[4868]: I1201 17:46:15.165298 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-z2bdk"] Dec 01 17:46:16 crc kubenswrapper[4868]: I1201 17:46:16.185488 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19e9c57b-8750-414f-b95b-3938c1356ed0" path="/var/lib/kubelet/pods/19e9c57b-8750-414f-b95b-3938c1356ed0/volumes" Dec 01 17:46:18 crc kubenswrapper[4868]: I1201 17:46:18.148555 4868 generic.go:334] "Generic (PLEG): container finished" podID="491393ef-ffbb-4a81-a477-5856765b7a48" containerID="98997845d571b01016aa693a291d1d21d2117a71f509381ce801cc152937438a" exitCode=0 Dec 01 17:46:18 crc kubenswrapper[4868]: I1201 17:46:18.148777 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pjj7m" event={"ID":"491393ef-ffbb-4a81-a477-5856765b7a48","Type":"ContainerDied","Data":"98997845d571b01016aa693a291d1d21d2117a71f509381ce801cc152937438a"} Dec 01 17:46:19 crc kubenswrapper[4868]: I1201 17:46:19.031828 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-bccf8f775-z2bdk" podUID="19e9c57b-8750-414f-b95b-3938c1356ed0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.189:5353: i/o timeout" Dec 01 17:46:19 crc kubenswrapper[4868]: I1201 17:46:19.541256 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:19 crc kubenswrapper[4868]: I1201 17:46:19.601576 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-scripts\") pod \"491393ef-ffbb-4a81-a477-5856765b7a48\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " Dec 01 17:46:19 crc kubenswrapper[4868]: I1201 17:46:19.601790 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-combined-ca-bundle\") pod \"491393ef-ffbb-4a81-a477-5856765b7a48\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " Dec 01 17:46:19 crc kubenswrapper[4868]: I1201 17:46:19.601845 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfr24\" (UniqueName: \"kubernetes.io/projected/491393ef-ffbb-4a81-a477-5856765b7a48-kube-api-access-dfr24\") pod \"491393ef-ffbb-4a81-a477-5856765b7a48\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " Dec 01 17:46:19 crc kubenswrapper[4868]: I1201 17:46:19.601868 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-config-data\") pod \"491393ef-ffbb-4a81-a477-5856765b7a48\" (UID: \"491393ef-ffbb-4a81-a477-5856765b7a48\") " Dec 01 17:46:19 crc kubenswrapper[4868]: I1201 17:46:19.607705 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-scripts" (OuterVolumeSpecName: "scripts") pod "491393ef-ffbb-4a81-a477-5856765b7a48" (UID: "491393ef-ffbb-4a81-a477-5856765b7a48"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:19 crc kubenswrapper[4868]: I1201 17:46:19.612231 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/491393ef-ffbb-4a81-a477-5856765b7a48-kube-api-access-dfr24" (OuterVolumeSpecName: "kube-api-access-dfr24") pod "491393ef-ffbb-4a81-a477-5856765b7a48" (UID: "491393ef-ffbb-4a81-a477-5856765b7a48"). InnerVolumeSpecName "kube-api-access-dfr24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:46:19 crc kubenswrapper[4868]: I1201 17:46:19.637072 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "491393ef-ffbb-4a81-a477-5856765b7a48" (UID: "491393ef-ffbb-4a81-a477-5856765b7a48"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:19 crc kubenswrapper[4868]: I1201 17:46:19.638152 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-config-data" (OuterVolumeSpecName: "config-data") pod "491393ef-ffbb-4a81-a477-5856765b7a48" (UID: "491393ef-ffbb-4a81-a477-5856765b7a48"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:19 crc kubenswrapper[4868]: I1201 17:46:19.704552 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:19 crc kubenswrapper[4868]: I1201 17:46:19.704594 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfr24\" (UniqueName: \"kubernetes.io/projected/491393ef-ffbb-4a81-a477-5856765b7a48-kube-api-access-dfr24\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:19 crc kubenswrapper[4868]: I1201 17:46:19.704609 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:19 crc kubenswrapper[4868]: I1201 17:46:19.704618 4868 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/491393ef-ffbb-4a81-a477-5856765b7a48-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:20 crc kubenswrapper[4868]: I1201 17:46:20.172368 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pjj7m" Dec 01 17:46:20 crc kubenswrapper[4868]: I1201 17:46:20.181366 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pjj7m" event={"ID":"491393ef-ffbb-4a81-a477-5856765b7a48","Type":"ContainerDied","Data":"110bad1c3880de1db14729d69a4fa7590d3266de7d6265dbb41b991840d635f6"} Dec 01 17:46:20 crc kubenswrapper[4868]: I1201 17:46:20.181408 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="110bad1c3880de1db14729d69a4fa7590d3266de7d6265dbb41b991840d635f6" Dec 01 17:46:20 crc kubenswrapper[4868]: I1201 17:46:20.351018 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:46:20 crc kubenswrapper[4868]: I1201 17:46:20.351311 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0eb5aa87-3260-4580-8827-b6a740ff01eb" containerName="nova-scheduler-scheduler" containerID="cri-o://ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482" gracePeriod=30 Dec 01 17:46:20 crc kubenswrapper[4868]: I1201 17:46:20.362302 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:46:20 crc kubenswrapper[4868]: I1201 17:46:20.362538 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ccced3dc-1af6-4959-a40f-c4f0dd654fca" containerName="nova-api-log" containerID="cri-o://0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b" gracePeriod=30 Dec 01 17:46:20 crc kubenswrapper[4868]: I1201 17:46:20.362645 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ccced3dc-1af6-4959-a40f-c4f0dd654fca" containerName="nova-api-api" containerID="cri-o://b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11" gracePeriod=30 Dec 01 17:46:20 crc kubenswrapper[4868]: I1201 17:46:20.434854 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:46:20 crc kubenswrapper[4868]: I1201 17:46:20.435157 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3239cfe0-6ea9-4580-b350-0ebeb56e2709" containerName="nova-metadata-log" containerID="cri-o://a986a0fbad2a9997257c0141ad8464f3d6acf24543c9f22bda40ef0b0ee579bc" gracePeriod=30 Dec 01 17:46:20 crc kubenswrapper[4868]: I1201 17:46:20.435219 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3239cfe0-6ea9-4580-b350-0ebeb56e2709" containerName="nova-metadata-metadata" containerID="cri-o://27b00cadf31aa7b4fe177daddbe129314acbd395f8ed1c08a61339775c6bd911" gracePeriod=30 Dec 01 17:46:20 crc kubenswrapper[4868]: I1201 17:46:20.953637 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.035633 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-internal-tls-certs\") pod \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.035772 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-config-data\") pod \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.035801 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-combined-ca-bundle\") pod \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.035850 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccced3dc-1af6-4959-a40f-c4f0dd654fca-logs\") pod \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.035890 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bw52\" (UniqueName: \"kubernetes.io/projected/ccced3dc-1af6-4959-a40f-c4f0dd654fca-kube-api-access-7bw52\") pod \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.036028 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-public-tls-certs\") pod \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.072116 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccced3dc-1af6-4959-a40f-c4f0dd654fca-logs" (OuterVolumeSpecName: "logs") pod "ccced3dc-1af6-4959-a40f-c4f0dd654fca" (UID: "ccced3dc-1af6-4959-a40f-c4f0dd654fca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.101149 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccced3dc-1af6-4959-a40f-c4f0dd654fca-kube-api-access-7bw52" (OuterVolumeSpecName: "kube-api-access-7bw52") pod "ccced3dc-1af6-4959-a40f-c4f0dd654fca" (UID: "ccced3dc-1af6-4959-a40f-c4f0dd654fca"). InnerVolumeSpecName "kube-api-access-7bw52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.111121 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-config-data" (OuterVolumeSpecName: "config-data") pod "ccced3dc-1af6-4959-a40f-c4f0dd654fca" (UID: "ccced3dc-1af6-4959-a40f-c4f0dd654fca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.113180 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ccced3dc-1af6-4959-a40f-c4f0dd654fca" (UID: "ccced3dc-1af6-4959-a40f-c4f0dd654fca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:21 crc kubenswrapper[4868]: E1201 17:46:21.132260 4868 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-public-tls-certs podName:ccced3dc-1af6-4959-a40f-c4f0dd654fca nodeName:}" failed. No retries permitted until 2025-12-01 17:46:21.632232615 +0000 UTC m=+1254.003343036 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "public-tls-certs" (UniqueName: "kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-public-tls-certs") pod "ccced3dc-1af6-4959-a40f-c4f0dd654fca" (UID: "ccced3dc-1af6-4959-a40f-c4f0dd654fca") : error deleting /var/lib/kubelet/pods/ccced3dc-1af6-4959-a40f-c4f0dd654fca/volume-subpaths: remove /var/lib/kubelet/pods/ccced3dc-1af6-4959-a40f-c4f0dd654fca/volume-subpaths: no such file or directory Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.136086 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ccced3dc-1af6-4959-a40f-c4f0dd654fca" (UID: "ccced3dc-1af6-4959-a40f-c4f0dd654fca"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.138311 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bw52\" (UniqueName: \"kubernetes.io/projected/ccced3dc-1af6-4959-a40f-c4f0dd654fca-kube-api-access-7bw52\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.138372 4868 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.138387 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.138399 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.138411 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccced3dc-1af6-4959-a40f-c4f0dd654fca-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.184924 4868 generic.go:334] "Generic (PLEG): container finished" podID="3239cfe0-6ea9-4580-b350-0ebeb56e2709" containerID="a986a0fbad2a9997257c0141ad8464f3d6acf24543c9f22bda40ef0b0ee579bc" exitCode=143 Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.184986 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3239cfe0-6ea9-4580-b350-0ebeb56e2709","Type":"ContainerDied","Data":"a986a0fbad2a9997257c0141ad8464f3d6acf24543c9f22bda40ef0b0ee579bc"} Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.186752 4868 generic.go:334] "Generic (PLEG): container finished" podID="ccced3dc-1af6-4959-a40f-c4f0dd654fca" containerID="b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11" exitCode=0 Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.186775 4868 generic.go:334] "Generic (PLEG): container finished" podID="ccced3dc-1af6-4959-a40f-c4f0dd654fca" containerID="0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b" exitCode=143 Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.186798 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ccced3dc-1af6-4959-a40f-c4f0dd654fca","Type":"ContainerDied","Data":"b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11"} Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.186825 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ccced3dc-1af6-4959-a40f-c4f0dd654fca","Type":"ContainerDied","Data":"0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b"} Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.186835 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ccced3dc-1af6-4959-a40f-c4f0dd654fca","Type":"ContainerDied","Data":"53b9168409f25ad2a933f73a3ee238813660706197252cda8925fc89ef5e630b"} Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.186844 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.186861 4868 scope.go:117] "RemoveContainer" containerID="b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.223957 4868 scope.go:117] "RemoveContainer" containerID="0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.253667 4868 scope.go:117] "RemoveContainer" containerID="b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11" Dec 01 17:46:21 crc kubenswrapper[4868]: E1201 17:46:21.254061 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11\": container with ID starting with b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11 not found: ID does not exist" containerID="b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.254092 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11"} err="failed to get container status \"b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11\": rpc error: code = NotFound desc = could not find container \"b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11\": container with ID starting with b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11 not found: ID does not exist" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.254115 4868 scope.go:117] "RemoveContainer" containerID="0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b" Dec 01 17:46:21 crc kubenswrapper[4868]: E1201 17:46:21.254327 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b\": container with ID starting with 0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b not found: ID does not exist" containerID="0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.254349 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b"} err="failed to get container status \"0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b\": rpc error: code = NotFound desc = could not find container \"0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b\": container with ID starting with 0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b not found: ID does not exist" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.254363 4868 scope.go:117] "RemoveContainer" containerID="b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.254545 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11"} err="failed to get container status \"b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11\": rpc error: code = NotFound desc = could not find container \"b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11\": container with ID starting with b0c6db0f77618d17281a8e3dc9de1fe5ca22773167bb4fc91a4849091123ce11 not found: ID does not exist" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.254563 4868 scope.go:117] "RemoveContainer" containerID="0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.254744 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b"} err="failed to get container status \"0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b\": rpc error: code = NotFound desc = could not find container \"0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b\": container with ID starting with 0e340055b2b7d2d8937b1603972ebda1af6dbe575e6310b5f333bcd7df5b313b not found: ID does not exist" Dec 01 17:46:21 crc kubenswrapper[4868]: E1201 17:46:21.410182 4868 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 17:46:21 crc kubenswrapper[4868]: E1201 17:46:21.411496 4868 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 17:46:21 crc kubenswrapper[4868]: E1201 17:46:21.413475 4868 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 17:46:21 crc kubenswrapper[4868]: E1201 17:46:21.413507 4868 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="0eb5aa87-3260-4580-8827-b6a740ff01eb" containerName="nova-scheduler-scheduler" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.649204 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-public-tls-certs\") pod \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\" (UID: \"ccced3dc-1af6-4959-a40f-c4f0dd654fca\") " Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.654097 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ccced3dc-1af6-4959-a40f-c4f0dd654fca" (UID: "ccced3dc-1af6-4959-a40f-c4f0dd654fca"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.751889 4868 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccced3dc-1af6-4959-a40f-c4f0dd654fca-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.824305 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.834189 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.856204 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 17:46:21 crc kubenswrapper[4868]: E1201 17:46:21.856657 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccced3dc-1af6-4959-a40f-c4f0dd654fca" containerName="nova-api-log" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.856678 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccced3dc-1af6-4959-a40f-c4f0dd654fca" containerName="nova-api-log" Dec 01 17:46:21 crc kubenswrapper[4868]: E1201 17:46:21.856697 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19e9c57b-8750-414f-b95b-3938c1356ed0" containerName="init" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.856705 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="19e9c57b-8750-414f-b95b-3938c1356ed0" containerName="init" Dec 01 17:46:21 crc kubenswrapper[4868]: E1201 17:46:21.856713 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19e9c57b-8750-414f-b95b-3938c1356ed0" containerName="dnsmasq-dns" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.856719 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="19e9c57b-8750-414f-b95b-3938c1356ed0" containerName="dnsmasq-dns" Dec 01 17:46:21 crc kubenswrapper[4868]: E1201 17:46:21.856739 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="491393ef-ffbb-4a81-a477-5856765b7a48" containerName="nova-manage" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.856744 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="491393ef-ffbb-4a81-a477-5856765b7a48" containerName="nova-manage" Dec 01 17:46:21 crc kubenswrapper[4868]: E1201 17:46:21.856759 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccced3dc-1af6-4959-a40f-c4f0dd654fca" containerName="nova-api-api" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.856765 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccced3dc-1af6-4959-a40f-c4f0dd654fca" containerName="nova-api-api" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.857017 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccced3dc-1af6-4959-a40f-c4f0dd654fca" containerName="nova-api-log" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.857037 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccced3dc-1af6-4959-a40f-c4f0dd654fca" containerName="nova-api-api" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.857049 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="19e9c57b-8750-414f-b95b-3938c1356ed0" containerName="dnsmasq-dns" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.857062 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="491393ef-ffbb-4a81-a477-5856765b7a48" containerName="nova-manage" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.858093 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.861692 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.861978 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.862205 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.868639 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.955115 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85lrx\" (UniqueName: \"kubernetes.io/projected/47392f20-a05b-44d5-be29-1d7588516c0a-kube-api-access-85lrx\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.955169 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47392f20-a05b-44d5-be29-1d7588516c0a-logs\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.955397 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47392f20-a05b-44d5-be29-1d7588516c0a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.955489 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47392f20-a05b-44d5-be29-1d7588516c0a-config-data\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.955524 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/47392f20-a05b-44d5-be29-1d7588516c0a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:21 crc kubenswrapper[4868]: I1201 17:46:21.955653 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47392f20-a05b-44d5-be29-1d7588516c0a-public-tls-certs\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:22 crc kubenswrapper[4868]: I1201 17:46:22.057778 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47392f20-a05b-44d5-be29-1d7588516c0a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:22 crc kubenswrapper[4868]: I1201 17:46:22.058167 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47392f20-a05b-44d5-be29-1d7588516c0a-config-data\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:22 crc kubenswrapper[4868]: I1201 17:46:22.058198 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/47392f20-a05b-44d5-be29-1d7588516c0a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:22 crc kubenswrapper[4868]: I1201 17:46:22.058295 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47392f20-a05b-44d5-be29-1d7588516c0a-public-tls-certs\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:22 crc kubenswrapper[4868]: I1201 17:46:22.058353 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85lrx\" (UniqueName: \"kubernetes.io/projected/47392f20-a05b-44d5-be29-1d7588516c0a-kube-api-access-85lrx\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:22 crc kubenswrapper[4868]: I1201 17:46:22.058388 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47392f20-a05b-44d5-be29-1d7588516c0a-logs\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:22 crc kubenswrapper[4868]: I1201 17:46:22.058974 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47392f20-a05b-44d5-be29-1d7588516c0a-logs\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:22 crc kubenswrapper[4868]: I1201 17:46:22.061492 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47392f20-a05b-44d5-be29-1d7588516c0a-public-tls-certs\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:22 crc kubenswrapper[4868]: I1201 17:46:22.061538 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47392f20-a05b-44d5-be29-1d7588516c0a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:22 crc kubenswrapper[4868]: I1201 17:46:22.061643 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47392f20-a05b-44d5-be29-1d7588516c0a-config-data\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:22 crc kubenswrapper[4868]: I1201 17:46:22.061935 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/47392f20-a05b-44d5-be29-1d7588516c0a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:22 crc kubenswrapper[4868]: I1201 17:46:22.085319 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85lrx\" (UniqueName: \"kubernetes.io/projected/47392f20-a05b-44d5-be29-1d7588516c0a-kube-api-access-85lrx\") pod \"nova-api-0\" (UID: \"47392f20-a05b-44d5-be29-1d7588516c0a\") " pod="openstack/nova-api-0" Dec 01 17:46:22 crc kubenswrapper[4868]: I1201 17:46:22.182723 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccced3dc-1af6-4959-a40f-c4f0dd654fca" path="/var/lib/kubelet/pods/ccced3dc-1af6-4959-a40f-c4f0dd654fca/volumes" Dec 01 17:46:22 crc kubenswrapper[4868]: I1201 17:46:22.184810 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:46:22 crc kubenswrapper[4868]: I1201 17:46:22.700658 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:46:23 crc kubenswrapper[4868]: I1201 17:46:23.208512 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"47392f20-a05b-44d5-be29-1d7588516c0a","Type":"ContainerStarted","Data":"80ff31cb03f7e230939e4cfbdf5a47189cb143a55c55d2e14351063c03d6917d"} Dec 01 17:46:23 crc kubenswrapper[4868]: I1201 17:46:23.208556 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"47392f20-a05b-44d5-be29-1d7588516c0a","Type":"ContainerStarted","Data":"3ca396012db4c812d22f5ccc9d80e5406a97cdec0355e69a8abeb8e706c0a898"} Dec 01 17:46:23 crc kubenswrapper[4868]: I1201 17:46:23.208567 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"47392f20-a05b-44d5-be29-1d7588516c0a","Type":"ContainerStarted","Data":"53bb07d29bb4ca69cbcd2572abdc5cc98e5cdcbc9274353498c22e8506413ab8"} Dec 01 17:46:23 crc kubenswrapper[4868]: I1201 17:46:23.235067 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.235045856 podStartE2EDuration="2.235045856s" podCreationTimestamp="2025-12-01 17:46:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:46:23.226906949 +0000 UTC m=+1255.598017370" watchObservedRunningTime="2025-12-01 17:46:23.235045856 +0000 UTC m=+1255.606156277" Dec 01 17:46:23 crc kubenswrapper[4868]: I1201 17:46:23.562287 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3239cfe0-6ea9-4580-b350-0ebeb56e2709" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": read tcp 10.217.0.2:45346->10.217.0.193:8775: read: connection reset by peer" Dec 01 17:46:23 crc kubenswrapper[4868]: I1201 17:46:23.562316 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3239cfe0-6ea9-4580-b350-0ebeb56e2709" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": read tcp 10.217.0.2:45344->10.217.0.193:8775: read: connection reset by peer" Dec 01 17:46:23 crc kubenswrapper[4868]: I1201 17:46:23.965805 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:46:23 crc kubenswrapper[4868]: I1201 17:46:23.989864 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2q75c\" (UniqueName: \"kubernetes.io/projected/3239cfe0-6ea9-4580-b350-0ebeb56e2709-kube-api-access-2q75c\") pod \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " Dec 01 17:46:23 crc kubenswrapper[4868]: I1201 17:46:23.989980 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-nova-metadata-tls-certs\") pod \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " Dec 01 17:46:23 crc kubenswrapper[4868]: I1201 17:46:23.990042 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3239cfe0-6ea9-4580-b350-0ebeb56e2709-logs\") pod \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " Dec 01 17:46:23 crc kubenswrapper[4868]: I1201 17:46:23.990079 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-config-data\") pod \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " Dec 01 17:46:23 crc kubenswrapper[4868]: I1201 17:46:23.990110 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-combined-ca-bundle\") pod \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\" (UID: \"3239cfe0-6ea9-4580-b350-0ebeb56e2709\") " Dec 01 17:46:23 crc kubenswrapper[4868]: I1201 17:46:23.993007 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3239cfe0-6ea9-4580-b350-0ebeb56e2709-logs" (OuterVolumeSpecName: "logs") pod "3239cfe0-6ea9-4580-b350-0ebeb56e2709" (UID: "3239cfe0-6ea9-4580-b350-0ebeb56e2709"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.022624 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3239cfe0-6ea9-4580-b350-0ebeb56e2709-kube-api-access-2q75c" (OuterVolumeSpecName: "kube-api-access-2q75c") pod "3239cfe0-6ea9-4580-b350-0ebeb56e2709" (UID: "3239cfe0-6ea9-4580-b350-0ebeb56e2709"). InnerVolumeSpecName "kube-api-access-2q75c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.054580 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-config-data" (OuterVolumeSpecName: "config-data") pod "3239cfe0-6ea9-4580-b350-0ebeb56e2709" (UID: "3239cfe0-6ea9-4580-b350-0ebeb56e2709"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.059805 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3239cfe0-6ea9-4580-b350-0ebeb56e2709" (UID: "3239cfe0-6ea9-4580-b350-0ebeb56e2709"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.085112 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "3239cfe0-6ea9-4580-b350-0ebeb56e2709" (UID: "3239cfe0-6ea9-4580-b350-0ebeb56e2709"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.091920 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2q75c\" (UniqueName: \"kubernetes.io/projected/3239cfe0-6ea9-4580-b350-0ebeb56e2709-kube-api-access-2q75c\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.091957 4868 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.091966 4868 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3239cfe0-6ea9-4580-b350-0ebeb56e2709-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.091977 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.091987 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3239cfe0-6ea9-4580-b350-0ebeb56e2709-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.219969 4868 generic.go:334] "Generic (PLEG): container finished" podID="3239cfe0-6ea9-4580-b350-0ebeb56e2709" containerID="27b00cadf31aa7b4fe177daddbe129314acbd395f8ed1c08a61339775c6bd911" exitCode=0 Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.220909 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.221393 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3239cfe0-6ea9-4580-b350-0ebeb56e2709","Type":"ContainerDied","Data":"27b00cadf31aa7b4fe177daddbe129314acbd395f8ed1c08a61339775c6bd911"} Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.221421 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3239cfe0-6ea9-4580-b350-0ebeb56e2709","Type":"ContainerDied","Data":"327f2cd2dbabbd3643cfd0ce04fe58d6e901b7439d55718561c251d7b0dc8017"} Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.221438 4868 scope.go:117] "RemoveContainer" containerID="27b00cadf31aa7b4fe177daddbe129314acbd395f8ed1c08a61339775c6bd911" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.245467 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.260015 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.264310 4868 scope.go:117] "RemoveContainer" containerID="a986a0fbad2a9997257c0141ad8464f3d6acf24543c9f22bda40ef0b0ee579bc" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.270127 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:46:24 crc kubenswrapper[4868]: E1201 17:46:24.270789 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3239cfe0-6ea9-4580-b350-0ebeb56e2709" containerName="nova-metadata-log" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.270815 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="3239cfe0-6ea9-4580-b350-0ebeb56e2709" containerName="nova-metadata-log" Dec 01 17:46:24 crc kubenswrapper[4868]: E1201 17:46:24.270866 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3239cfe0-6ea9-4580-b350-0ebeb56e2709" containerName="nova-metadata-metadata" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.270875 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="3239cfe0-6ea9-4580-b350-0ebeb56e2709" containerName="nova-metadata-metadata" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.271117 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="3239cfe0-6ea9-4580-b350-0ebeb56e2709" containerName="nova-metadata-log" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.271150 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="3239cfe0-6ea9-4580-b350-0ebeb56e2709" containerName="nova-metadata-metadata" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.272767 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.275878 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.276272 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.288961 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.301555 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/76578957-dead-4612-9d4d-bd9d19c589ef-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"76578957-dead-4612-9d4d-bd9d19c589ef\") " pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.301623 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76578957-dead-4612-9d4d-bd9d19c589ef-logs\") pod \"nova-metadata-0\" (UID: \"76578957-dead-4612-9d4d-bd9d19c589ef\") " pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.301660 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76578957-dead-4612-9d4d-bd9d19c589ef-config-data\") pod \"nova-metadata-0\" (UID: \"76578957-dead-4612-9d4d-bd9d19c589ef\") " pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.301857 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76578957-dead-4612-9d4d-bd9d19c589ef-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"76578957-dead-4612-9d4d-bd9d19c589ef\") " pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.301893 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sbbd\" (UniqueName: \"kubernetes.io/projected/76578957-dead-4612-9d4d-bd9d19c589ef-kube-api-access-7sbbd\") pod \"nova-metadata-0\" (UID: \"76578957-dead-4612-9d4d-bd9d19c589ef\") " pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.314887 4868 scope.go:117] "RemoveContainer" containerID="27b00cadf31aa7b4fe177daddbe129314acbd395f8ed1c08a61339775c6bd911" Dec 01 17:46:24 crc kubenswrapper[4868]: E1201 17:46:24.315346 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27b00cadf31aa7b4fe177daddbe129314acbd395f8ed1c08a61339775c6bd911\": container with ID starting with 27b00cadf31aa7b4fe177daddbe129314acbd395f8ed1c08a61339775c6bd911 not found: ID does not exist" containerID="27b00cadf31aa7b4fe177daddbe129314acbd395f8ed1c08a61339775c6bd911" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.315376 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27b00cadf31aa7b4fe177daddbe129314acbd395f8ed1c08a61339775c6bd911"} err="failed to get container status \"27b00cadf31aa7b4fe177daddbe129314acbd395f8ed1c08a61339775c6bd911\": rpc error: code = NotFound desc = could not find container \"27b00cadf31aa7b4fe177daddbe129314acbd395f8ed1c08a61339775c6bd911\": container with ID starting with 27b00cadf31aa7b4fe177daddbe129314acbd395f8ed1c08a61339775c6bd911 not found: ID does not exist" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.315399 4868 scope.go:117] "RemoveContainer" containerID="a986a0fbad2a9997257c0141ad8464f3d6acf24543c9f22bda40ef0b0ee579bc" Dec 01 17:46:24 crc kubenswrapper[4868]: E1201 17:46:24.315819 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a986a0fbad2a9997257c0141ad8464f3d6acf24543c9f22bda40ef0b0ee579bc\": container with ID starting with a986a0fbad2a9997257c0141ad8464f3d6acf24543c9f22bda40ef0b0ee579bc not found: ID does not exist" containerID="a986a0fbad2a9997257c0141ad8464f3d6acf24543c9f22bda40ef0b0ee579bc" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.315869 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a986a0fbad2a9997257c0141ad8464f3d6acf24543c9f22bda40ef0b0ee579bc"} err="failed to get container status \"a986a0fbad2a9997257c0141ad8464f3d6acf24543c9f22bda40ef0b0ee579bc\": rpc error: code = NotFound desc = could not find container \"a986a0fbad2a9997257c0141ad8464f3d6acf24543c9f22bda40ef0b0ee579bc\": container with ID starting with a986a0fbad2a9997257c0141ad8464f3d6acf24543c9f22bda40ef0b0ee579bc not found: ID does not exist" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.403201 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/76578957-dead-4612-9d4d-bd9d19c589ef-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"76578957-dead-4612-9d4d-bd9d19c589ef\") " pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.403273 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76578957-dead-4612-9d4d-bd9d19c589ef-logs\") pod \"nova-metadata-0\" (UID: \"76578957-dead-4612-9d4d-bd9d19c589ef\") " pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.403302 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76578957-dead-4612-9d4d-bd9d19c589ef-config-data\") pod \"nova-metadata-0\" (UID: \"76578957-dead-4612-9d4d-bd9d19c589ef\") " pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.403356 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76578957-dead-4612-9d4d-bd9d19c589ef-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"76578957-dead-4612-9d4d-bd9d19c589ef\") " pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.403391 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sbbd\" (UniqueName: \"kubernetes.io/projected/76578957-dead-4612-9d4d-bd9d19c589ef-kube-api-access-7sbbd\") pod \"nova-metadata-0\" (UID: \"76578957-dead-4612-9d4d-bd9d19c589ef\") " pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.404898 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76578957-dead-4612-9d4d-bd9d19c589ef-logs\") pod \"nova-metadata-0\" (UID: \"76578957-dead-4612-9d4d-bd9d19c589ef\") " pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.411723 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76578957-dead-4612-9d4d-bd9d19c589ef-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"76578957-dead-4612-9d4d-bd9d19c589ef\") " pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.415929 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/76578957-dead-4612-9d4d-bd9d19c589ef-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"76578957-dead-4612-9d4d-bd9d19c589ef\") " pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.421550 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76578957-dead-4612-9d4d-bd9d19c589ef-config-data\") pod \"nova-metadata-0\" (UID: \"76578957-dead-4612-9d4d-bd9d19c589ef\") " pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.423484 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sbbd\" (UniqueName: \"kubernetes.io/projected/76578957-dead-4612-9d4d-bd9d19c589ef-kube-api-access-7sbbd\") pod \"nova-metadata-0\" (UID: \"76578957-dead-4612-9d4d-bd9d19c589ef\") " pod="openstack/nova-metadata-0" Dec 01 17:46:24 crc kubenswrapper[4868]: I1201 17:46:24.602536 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.210448 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:46:25 crc kubenswrapper[4868]: W1201 17:46:25.213094 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76578957_dead_4612_9d4d_bd9d19c589ef.slice/crio-9202f2812ba03f099628f3b5c52c00c7c8884923ff76707c1902a63e44ced12b WatchSource:0}: Error finding container 9202f2812ba03f099628f3b5c52c00c7c8884923ff76707c1902a63e44ced12b: Status 404 returned error can't find the container with id 9202f2812ba03f099628f3b5c52c00c7c8884923ff76707c1902a63e44ced12b Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.229189 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"76578957-dead-4612-9d4d-bd9d19c589ef","Type":"ContainerStarted","Data":"9202f2812ba03f099628f3b5c52c00c7c8884923ff76707c1902a63e44ced12b"} Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.781349 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.865703 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgvn5\" (UniqueName: \"kubernetes.io/projected/0eb5aa87-3260-4580-8827-b6a740ff01eb-kube-api-access-qgvn5\") pod \"0eb5aa87-3260-4580-8827-b6a740ff01eb\" (UID: \"0eb5aa87-3260-4580-8827-b6a740ff01eb\") " Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.866296 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0eb5aa87-3260-4580-8827-b6a740ff01eb-config-data\") pod \"0eb5aa87-3260-4580-8827-b6a740ff01eb\" (UID: \"0eb5aa87-3260-4580-8827-b6a740ff01eb\") " Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.866448 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0eb5aa87-3260-4580-8827-b6a740ff01eb-combined-ca-bundle\") pod \"0eb5aa87-3260-4580-8827-b6a740ff01eb\" (UID: \"0eb5aa87-3260-4580-8827-b6a740ff01eb\") " Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.880297 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0eb5aa87-3260-4580-8827-b6a740ff01eb-kube-api-access-qgvn5" (OuterVolumeSpecName: "kube-api-access-qgvn5") pod "0eb5aa87-3260-4580-8827-b6a740ff01eb" (UID: "0eb5aa87-3260-4580-8827-b6a740ff01eb"). InnerVolumeSpecName "kube-api-access-qgvn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.905235 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.905288 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.905332 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.905632 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eb5aa87-3260-4580-8827-b6a740ff01eb-config-data" (OuterVolumeSpecName: "config-data") pod "0eb5aa87-3260-4580-8827-b6a740ff01eb" (UID: "0eb5aa87-3260-4580-8827-b6a740ff01eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.906020 4868 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"35d0f81eb6527dda2c61338bf9c12106c16489b5c045a71ea474df12051c7650"} pod="openshift-machine-config-operator/machine-config-daemon-njgpn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.906071 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" containerID="cri-o://35d0f81eb6527dda2c61338bf9c12106c16489b5c045a71ea474df12051c7650" gracePeriod=600 Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.916132 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eb5aa87-3260-4580-8827-b6a740ff01eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0eb5aa87-3260-4580-8827-b6a740ff01eb" (UID: "0eb5aa87-3260-4580-8827-b6a740ff01eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.968764 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgvn5\" (UniqueName: \"kubernetes.io/projected/0eb5aa87-3260-4580-8827-b6a740ff01eb-kube-api-access-qgvn5\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.968815 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0eb5aa87-3260-4580-8827-b6a740ff01eb-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:25 crc kubenswrapper[4868]: I1201 17:46:25.968828 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0eb5aa87-3260-4580-8827-b6a740ff01eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.183391 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3239cfe0-6ea9-4580-b350-0ebeb56e2709" path="/var/lib/kubelet/pods/3239cfe0-6ea9-4580-b350-0ebeb56e2709/volumes" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.241097 4868 generic.go:334] "Generic (PLEG): container finished" podID="ec4e08bf-df56-4490-8e79-a4175c727405" containerID="35d0f81eb6527dda2c61338bf9c12106c16489b5c045a71ea474df12051c7650" exitCode=0 Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.241145 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerDied","Data":"35d0f81eb6527dda2c61338bf9c12106c16489b5c045a71ea474df12051c7650"} Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.241199 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"998e9f9a5d44e576ae1e608687cc6cddbcaf1995065246b296eae79bc764efdb"} Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.241216 4868 scope.go:117] "RemoveContainer" containerID="2d56a7ef7fc39bd11e02dfcafda9203c34f3593dd8178f744951427000e66a77" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.244499 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"76578957-dead-4612-9d4d-bd9d19c589ef","Type":"ContainerStarted","Data":"9003767f49027b65c673b1f42786e868e453b49f7586ce2ea9ae7f8857aa7677"} Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.244525 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"76578957-dead-4612-9d4d-bd9d19c589ef","Type":"ContainerStarted","Data":"8bbfc593112258ec7d91c10f6ac1ee34e8ed90993c6588a983610245c4ee6838"} Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.247425 4868 generic.go:334] "Generic (PLEG): container finished" podID="0eb5aa87-3260-4580-8827-b6a740ff01eb" containerID="ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482" exitCode=0 Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.247457 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0eb5aa87-3260-4580-8827-b6a740ff01eb","Type":"ContainerDied","Data":"ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482"} Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.247467 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.247474 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0eb5aa87-3260-4580-8827-b6a740ff01eb","Type":"ContainerDied","Data":"212689681e76de1303085bcd70599b720bf8469f60fed03bdb650e3e0816bfc6"} Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.267598 4868 scope.go:117] "RemoveContainer" containerID="ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.277290 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.295036 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.307125 4868 scope.go:117] "RemoveContainer" containerID="ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.311293 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:46:26 crc kubenswrapper[4868]: E1201 17:46:26.311676 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482\": container with ID starting with ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482 not found: ID does not exist" containerID="ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.311707 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482"} err="failed to get container status \"ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482\": rpc error: code = NotFound desc = could not find container \"ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482\": container with ID starting with ca284671fa2b5afa12809737def4e5a946d89392019fffc9eeb8346583363482 not found: ID does not exist" Dec 01 17:46:26 crc kubenswrapper[4868]: E1201 17:46:26.311745 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eb5aa87-3260-4580-8827-b6a740ff01eb" containerName="nova-scheduler-scheduler" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.311759 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eb5aa87-3260-4580-8827-b6a740ff01eb" containerName="nova-scheduler-scheduler" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.311995 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="0eb5aa87-3260-4580-8827-b6a740ff01eb" containerName="nova-scheduler-scheduler" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.312705 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.314117 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.314092761 podStartE2EDuration="2.314092761s" podCreationTimestamp="2025-12-01 17:46:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:46:26.295605773 +0000 UTC m=+1258.666716184" watchObservedRunningTime="2025-12-01 17:46:26.314092761 +0000 UTC m=+1258.685203172" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.315212 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.347054 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.377423 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea59e84e-774b-4d8f-af90-1de84fe578d7-config-data\") pod \"nova-scheduler-0\" (UID: \"ea59e84e-774b-4d8f-af90-1de84fe578d7\") " pod="openstack/nova-scheduler-0" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.377581 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea59e84e-774b-4d8f-af90-1de84fe578d7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ea59e84e-774b-4d8f-af90-1de84fe578d7\") " pod="openstack/nova-scheduler-0" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.377804 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzxw6\" (UniqueName: \"kubernetes.io/projected/ea59e84e-774b-4d8f-af90-1de84fe578d7-kube-api-access-vzxw6\") pod \"nova-scheduler-0\" (UID: \"ea59e84e-774b-4d8f-af90-1de84fe578d7\") " pod="openstack/nova-scheduler-0" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.479871 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzxw6\" (UniqueName: \"kubernetes.io/projected/ea59e84e-774b-4d8f-af90-1de84fe578d7-kube-api-access-vzxw6\") pod \"nova-scheduler-0\" (UID: \"ea59e84e-774b-4d8f-af90-1de84fe578d7\") " pod="openstack/nova-scheduler-0" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.480031 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea59e84e-774b-4d8f-af90-1de84fe578d7-config-data\") pod \"nova-scheduler-0\" (UID: \"ea59e84e-774b-4d8f-af90-1de84fe578d7\") " pod="openstack/nova-scheduler-0" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.480132 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea59e84e-774b-4d8f-af90-1de84fe578d7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ea59e84e-774b-4d8f-af90-1de84fe578d7\") " pod="openstack/nova-scheduler-0" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.485980 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea59e84e-774b-4d8f-af90-1de84fe578d7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ea59e84e-774b-4d8f-af90-1de84fe578d7\") " pod="openstack/nova-scheduler-0" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.486533 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea59e84e-774b-4d8f-af90-1de84fe578d7-config-data\") pod \"nova-scheduler-0\" (UID: \"ea59e84e-774b-4d8f-af90-1de84fe578d7\") " pod="openstack/nova-scheduler-0" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.500649 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzxw6\" (UniqueName: \"kubernetes.io/projected/ea59e84e-774b-4d8f-af90-1de84fe578d7-kube-api-access-vzxw6\") pod \"nova-scheduler-0\" (UID: \"ea59e84e-774b-4d8f-af90-1de84fe578d7\") " pod="openstack/nova-scheduler-0" Dec 01 17:46:26 crc kubenswrapper[4868]: I1201 17:46:26.636302 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 17:46:27 crc kubenswrapper[4868]: I1201 17:46:27.065995 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:46:27 crc kubenswrapper[4868]: W1201 17:46:27.066780 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea59e84e_774b_4d8f_af90_1de84fe578d7.slice/crio-d963e9b23e5fbaeabddcfe83de0669bdbe77e9b7ff6aaea439176545538394d6 WatchSource:0}: Error finding container d963e9b23e5fbaeabddcfe83de0669bdbe77e9b7ff6aaea439176545538394d6: Status 404 returned error can't find the container with id d963e9b23e5fbaeabddcfe83de0669bdbe77e9b7ff6aaea439176545538394d6 Dec 01 17:46:27 crc kubenswrapper[4868]: I1201 17:46:27.261788 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ea59e84e-774b-4d8f-af90-1de84fe578d7","Type":"ContainerStarted","Data":"1902d3e4721ddce5b61f35bfa8289b487c28300216517c8659113df4ff9cecff"} Dec 01 17:46:27 crc kubenswrapper[4868]: I1201 17:46:27.261825 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ea59e84e-774b-4d8f-af90-1de84fe578d7","Type":"ContainerStarted","Data":"d963e9b23e5fbaeabddcfe83de0669bdbe77e9b7ff6aaea439176545538394d6"} Dec 01 17:46:28 crc kubenswrapper[4868]: I1201 17:46:28.208412 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0eb5aa87-3260-4580-8827-b6a740ff01eb" path="/var/lib/kubelet/pods/0eb5aa87-3260-4580-8827-b6a740ff01eb/volumes" Dec 01 17:46:29 crc kubenswrapper[4868]: I1201 17:46:29.603116 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 17:46:29 crc kubenswrapper[4868]: I1201 17:46:29.603448 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 17:46:31 crc kubenswrapper[4868]: I1201 17:46:31.636992 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 17:46:32 crc kubenswrapper[4868]: I1201 17:46:32.187554 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 17:46:32 crc kubenswrapper[4868]: I1201 17:46:32.187615 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 17:46:33 crc kubenswrapper[4868]: I1201 17:46:33.203238 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="47392f20-a05b-44d5-be29-1d7588516c0a" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 17:46:33 crc kubenswrapper[4868]: I1201 17:46:33.203278 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="47392f20-a05b-44d5-be29-1d7588516c0a" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 17:46:34 crc kubenswrapper[4868]: I1201 17:46:34.603199 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 17:46:34 crc kubenswrapper[4868]: I1201 17:46:34.603361 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 17:46:35 crc kubenswrapper[4868]: I1201 17:46:35.615118 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="76578957-dead-4612-9d4d-bd9d19c589ef" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 17:46:35 crc kubenswrapper[4868]: I1201 17:46:35.616030 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="76578957-dead-4612-9d4d-bd9d19c589ef" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 17:46:36 crc kubenswrapper[4868]: I1201 17:46:36.637276 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 17:46:36 crc kubenswrapper[4868]: I1201 17:46:36.662589 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 17:46:36 crc kubenswrapper[4868]: I1201 17:46:36.687741 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=10.687724022 podStartE2EDuration="10.687724022s" podCreationTimestamp="2025-12-01 17:46:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:46:27.283620345 +0000 UTC m=+1259.654730756" watchObservedRunningTime="2025-12-01 17:46:36.687724022 +0000 UTC m=+1269.058834433" Dec 01 17:46:37 crc kubenswrapper[4868]: I1201 17:46:37.399149 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 17:46:38 crc kubenswrapper[4868]: I1201 17:46:38.426347 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 17:46:42 crc kubenswrapper[4868]: I1201 17:46:42.190780 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 17:46:42 crc kubenswrapper[4868]: I1201 17:46:42.191570 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 17:46:42 crc kubenswrapper[4868]: I1201 17:46:42.193114 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 17:46:42 crc kubenswrapper[4868]: I1201 17:46:42.204823 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 17:46:42 crc kubenswrapper[4868]: I1201 17:46:42.427346 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 17:46:42 crc kubenswrapper[4868]: I1201 17:46:42.449762 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 17:46:44 crc kubenswrapper[4868]: I1201 17:46:44.608133 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 17:46:44 crc kubenswrapper[4868]: I1201 17:46:44.609906 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 17:46:44 crc kubenswrapper[4868]: I1201 17:46:44.612923 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 17:46:45 crc kubenswrapper[4868]: I1201 17:46:45.483604 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 17:46:53 crc kubenswrapper[4868]: I1201 17:46:53.303149 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 17:46:54 crc kubenswrapper[4868]: I1201 17:46:54.483582 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 17:46:57 crc kubenswrapper[4868]: I1201 17:46:57.908211 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="70aee7bf-d819-4355-a98c-4a106f3e6f16" containerName="rabbitmq" containerID="cri-o://a4829417dd1177e3ef6e9cbc5057cf402a287d114b27dd531e97f35d3beac596" gracePeriod=604796 Dec 01 17:46:59 crc kubenswrapper[4868]: I1201 17:46:59.086711 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="3ff4f488-06c8-4ceb-848a-4ba68b334752" containerName="rabbitmq" containerID="cri-o://a2b5914fe7354791d6773f33ac9378c19a8e5b4afca7990b0b44afa754192933" gracePeriod=604796 Dec 01 17:47:00 crc kubenswrapper[4868]: I1201 17:47:00.627633 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="70aee7bf-d819-4355-a98c-4a106f3e6f16" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Dec 01 17:47:00 crc kubenswrapper[4868]: I1201 17:47:00.976833 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="3ff4f488-06c8-4ceb-848a-4ba68b334752" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.490248 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.519187 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/70aee7bf-d819-4355-a98c-4a106f3e6f16-pod-info\") pod \"70aee7bf-d819-4355-a98c-4a106f3e6f16\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.519247 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgls7\" (UniqueName: \"kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-kube-api-access-pgls7\") pod \"70aee7bf-d819-4355-a98c-4a106f3e6f16\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.519282 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"70aee7bf-d819-4355-a98c-4a106f3e6f16\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.519321 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-tls\") pod \"70aee7bf-d819-4355-a98c-4a106f3e6f16\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.519418 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-server-conf\") pod \"70aee7bf-d819-4355-a98c-4a106f3e6f16\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.519442 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/70aee7bf-d819-4355-a98c-4a106f3e6f16-erlang-cookie-secret\") pod \"70aee7bf-d819-4355-a98c-4a106f3e6f16\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.519533 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-erlang-cookie\") pod \"70aee7bf-d819-4355-a98c-4a106f3e6f16\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.519606 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-config-data\") pod \"70aee7bf-d819-4355-a98c-4a106f3e6f16\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.519648 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-plugins-conf\") pod \"70aee7bf-d819-4355-a98c-4a106f3e6f16\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.519714 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-confd\") pod \"70aee7bf-d819-4355-a98c-4a106f3e6f16\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.519745 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-plugins\") pod \"70aee7bf-d819-4355-a98c-4a106f3e6f16\" (UID: \"70aee7bf-d819-4355-a98c-4a106f3e6f16\") " Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.520659 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "70aee7bf-d819-4355-a98c-4a106f3e6f16" (UID: "70aee7bf-d819-4355-a98c-4a106f3e6f16"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.528502 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "70aee7bf-d819-4355-a98c-4a106f3e6f16" (UID: "70aee7bf-d819-4355-a98c-4a106f3e6f16"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.529055 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "70aee7bf-d819-4355-a98c-4a106f3e6f16" (UID: "70aee7bf-d819-4355-a98c-4a106f3e6f16"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.529156 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "70aee7bf-d819-4355-a98c-4a106f3e6f16" (UID: "70aee7bf-d819-4355-a98c-4a106f3e6f16"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.531104 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70aee7bf-d819-4355-a98c-4a106f3e6f16-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "70aee7bf-d819-4355-a98c-4a106f3e6f16" (UID: "70aee7bf-d819-4355-a98c-4a106f3e6f16"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.541187 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/70aee7bf-d819-4355-a98c-4a106f3e6f16-pod-info" (OuterVolumeSpecName: "pod-info") pod "70aee7bf-d819-4355-a98c-4a106f3e6f16" (UID: "70aee7bf-d819-4355-a98c-4a106f3e6f16"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.541738 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-kube-api-access-pgls7" (OuterVolumeSpecName: "kube-api-access-pgls7") pod "70aee7bf-d819-4355-a98c-4a106f3e6f16" (UID: "70aee7bf-d819-4355-a98c-4a106f3e6f16"). InnerVolumeSpecName "kube-api-access-pgls7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.552797 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "70aee7bf-d819-4355-a98c-4a106f3e6f16" (UID: "70aee7bf-d819-4355-a98c-4a106f3e6f16"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.579060 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-config-data" (OuterVolumeSpecName: "config-data") pod "70aee7bf-d819-4355-a98c-4a106f3e6f16" (UID: "70aee7bf-d819-4355-a98c-4a106f3e6f16"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.623009 4868 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.623033 4868 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.623042 4868 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/70aee7bf-d819-4355-a98c-4a106f3e6f16-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.623051 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgls7\" (UniqueName: \"kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-kube-api-access-pgls7\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.623075 4868 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.623085 4868 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.623093 4868 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/70aee7bf-d819-4355-a98c-4a106f3e6f16-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.623102 4868 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.623112 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.640006 4868 generic.go:334] "Generic (PLEG): container finished" podID="70aee7bf-d819-4355-a98c-4a106f3e6f16" containerID="a4829417dd1177e3ef6e9cbc5057cf402a287d114b27dd531e97f35d3beac596" exitCode=0 Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.640051 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"70aee7bf-d819-4355-a98c-4a106f3e6f16","Type":"ContainerDied","Data":"a4829417dd1177e3ef6e9cbc5057cf402a287d114b27dd531e97f35d3beac596"} Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.640079 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"70aee7bf-d819-4355-a98c-4a106f3e6f16","Type":"ContainerDied","Data":"36fd322c034c02e8064e5407dbf69bd923bb67d3ba40e2c1eadad630a25cbed2"} Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.640096 4868 scope.go:117] "RemoveContainer" containerID="a4829417dd1177e3ef6e9cbc5057cf402a287d114b27dd531e97f35d3beac596" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.640218 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.644926 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-server-conf" (OuterVolumeSpecName: "server-conf") pod "70aee7bf-d819-4355-a98c-4a106f3e6f16" (UID: "70aee7bf-d819-4355-a98c-4a106f3e6f16"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.645351 4868 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.674622 4868 scope.go:117] "RemoveContainer" containerID="571a7ff19cbb3694810ea83e478a19bfa550fa8fb19697462b8c615086834487" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.714442 4868 scope.go:117] "RemoveContainer" containerID="a4829417dd1177e3ef6e9cbc5057cf402a287d114b27dd531e97f35d3beac596" Dec 01 17:47:04 crc kubenswrapper[4868]: E1201 17:47:04.715240 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4829417dd1177e3ef6e9cbc5057cf402a287d114b27dd531e97f35d3beac596\": container with ID starting with a4829417dd1177e3ef6e9cbc5057cf402a287d114b27dd531e97f35d3beac596 not found: ID does not exist" containerID="a4829417dd1177e3ef6e9cbc5057cf402a287d114b27dd531e97f35d3beac596" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.715292 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4829417dd1177e3ef6e9cbc5057cf402a287d114b27dd531e97f35d3beac596"} err="failed to get container status \"a4829417dd1177e3ef6e9cbc5057cf402a287d114b27dd531e97f35d3beac596\": rpc error: code = NotFound desc = could not find container \"a4829417dd1177e3ef6e9cbc5057cf402a287d114b27dd531e97f35d3beac596\": container with ID starting with a4829417dd1177e3ef6e9cbc5057cf402a287d114b27dd531e97f35d3beac596 not found: ID does not exist" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.715320 4868 scope.go:117] "RemoveContainer" containerID="571a7ff19cbb3694810ea83e478a19bfa550fa8fb19697462b8c615086834487" Dec 01 17:47:04 crc kubenswrapper[4868]: E1201 17:47:04.715596 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"571a7ff19cbb3694810ea83e478a19bfa550fa8fb19697462b8c615086834487\": container with ID starting with 571a7ff19cbb3694810ea83e478a19bfa550fa8fb19697462b8c615086834487 not found: ID does not exist" containerID="571a7ff19cbb3694810ea83e478a19bfa550fa8fb19697462b8c615086834487" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.715623 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"571a7ff19cbb3694810ea83e478a19bfa550fa8fb19697462b8c615086834487"} err="failed to get container status \"571a7ff19cbb3694810ea83e478a19bfa550fa8fb19697462b8c615086834487\": rpc error: code = NotFound desc = could not find container \"571a7ff19cbb3694810ea83e478a19bfa550fa8fb19697462b8c615086834487\": container with ID starting with 571a7ff19cbb3694810ea83e478a19bfa550fa8fb19697462b8c615086834487 not found: ID does not exist" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.724932 4868 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/70aee7bf-d819-4355-a98c-4a106f3e6f16-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.724995 4868 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.740936 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "70aee7bf-d819-4355-a98c-4a106f3e6f16" (UID: "70aee7bf-d819-4355-a98c-4a106f3e6f16"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.826862 4868 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/70aee7bf-d819-4355-a98c-4a106f3e6f16-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.973060 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 17:47:04 crc kubenswrapper[4868]: I1201 17:47:04.982155 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.006141 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 17:47:05 crc kubenswrapper[4868]: E1201 17:47:05.006533 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70aee7bf-d819-4355-a98c-4a106f3e6f16" containerName="setup-container" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.006568 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="70aee7bf-d819-4355-a98c-4a106f3e6f16" containerName="setup-container" Dec 01 17:47:05 crc kubenswrapper[4868]: E1201 17:47:05.006593 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70aee7bf-d819-4355-a98c-4a106f3e6f16" containerName="rabbitmq" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.006599 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="70aee7bf-d819-4355-a98c-4a106f3e6f16" containerName="rabbitmq" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.006784 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="70aee7bf-d819-4355-a98c-4a106f3e6f16" containerName="rabbitmq" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.007737 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.010366 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.010439 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.010459 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.010492 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-94wgt" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.010665 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.010668 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.011492 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.034754 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.036408 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b05cb910-a007-46d5-8182-eb1d30c5f953-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.036521 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b05cb910-a007-46d5-8182-eb1d30c5f953-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.036568 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b05cb910-a007-46d5-8182-eb1d30c5f953-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.036610 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.036670 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b05cb910-a007-46d5-8182-eb1d30c5f953-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.036845 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b05cb910-a007-46d5-8182-eb1d30c5f953-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.036886 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b05cb910-a007-46d5-8182-eb1d30c5f953-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.036979 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b05cb910-a007-46d5-8182-eb1d30c5f953-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.037074 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b05cb910-a007-46d5-8182-eb1d30c5f953-config-data\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.037111 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv725\" (UniqueName: \"kubernetes.io/projected/b05cb910-a007-46d5-8182-eb1d30c5f953-kube-api-access-gv725\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.037161 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b05cb910-a007-46d5-8182-eb1d30c5f953-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.139381 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b05cb910-a007-46d5-8182-eb1d30c5f953-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.139436 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b05cb910-a007-46d5-8182-eb1d30c5f953-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.139488 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b05cb910-a007-46d5-8182-eb1d30c5f953-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.139542 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b05cb910-a007-46d5-8182-eb1d30c5f953-config-data\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.139564 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv725\" (UniqueName: \"kubernetes.io/projected/b05cb910-a007-46d5-8182-eb1d30c5f953-kube-api-access-gv725\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.139593 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b05cb910-a007-46d5-8182-eb1d30c5f953-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.139657 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b05cb910-a007-46d5-8182-eb1d30c5f953-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.139710 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b05cb910-a007-46d5-8182-eb1d30c5f953-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.139739 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b05cb910-a007-46d5-8182-eb1d30c5f953-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.139760 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.139782 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b05cb910-a007-46d5-8182-eb1d30c5f953-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.140630 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b05cb910-a007-46d5-8182-eb1d30c5f953-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.141031 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b05cb910-a007-46d5-8182-eb1d30c5f953-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.141037 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b05cb910-a007-46d5-8182-eb1d30c5f953-config-data\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.141175 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.141366 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b05cb910-a007-46d5-8182-eb1d30c5f953-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.142782 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b05cb910-a007-46d5-8182-eb1d30c5f953-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.143581 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b05cb910-a007-46d5-8182-eb1d30c5f953-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.145236 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b05cb910-a007-46d5-8182-eb1d30c5f953-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.147425 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b05cb910-a007-46d5-8182-eb1d30c5f953-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.156484 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b05cb910-a007-46d5-8182-eb1d30c5f953-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.158191 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv725\" (UniqueName: \"kubernetes.io/projected/b05cb910-a007-46d5-8182-eb1d30c5f953-kube-api-access-gv725\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.190672 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"b05cb910-a007-46d5-8182-eb1d30c5f953\") " pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.342550 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.654044 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.655663 4868 generic.go:334] "Generic (PLEG): container finished" podID="3ff4f488-06c8-4ceb-848a-4ba68b334752" containerID="a2b5914fe7354791d6773f33ac9378c19a8e5b4afca7990b0b44afa754192933" exitCode=0 Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.655692 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3ff4f488-06c8-4ceb-848a-4ba68b334752","Type":"ContainerDied","Data":"a2b5914fe7354791d6773f33ac9378c19a8e5b4afca7990b0b44afa754192933"} Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.655714 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3ff4f488-06c8-4ceb-848a-4ba68b334752","Type":"ContainerDied","Data":"e98cc91c9e8fb7ee706b45be8b9b3fbbbfe6e4dcd75eb3638b68007528d4acda"} Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.655729 4868 scope.go:117] "RemoveContainer" containerID="a2b5914fe7354791d6773f33ac9378c19a8e5b4afca7990b0b44afa754192933" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.693446 4868 scope.go:117] "RemoveContainer" containerID="bbc47c78ecea10373d709f8016370bad805694e3d2800bf35760bd08eb36dc39" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.728238 4868 scope.go:117] "RemoveContainer" containerID="a2b5914fe7354791d6773f33ac9378c19a8e5b4afca7990b0b44afa754192933" Dec 01 17:47:05 crc kubenswrapper[4868]: E1201 17:47:05.728740 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2b5914fe7354791d6773f33ac9378c19a8e5b4afca7990b0b44afa754192933\": container with ID starting with a2b5914fe7354791d6773f33ac9378c19a8e5b4afca7990b0b44afa754192933 not found: ID does not exist" containerID="a2b5914fe7354791d6773f33ac9378c19a8e5b4afca7990b0b44afa754192933" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.728795 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2b5914fe7354791d6773f33ac9378c19a8e5b4afca7990b0b44afa754192933"} err="failed to get container status \"a2b5914fe7354791d6773f33ac9378c19a8e5b4afca7990b0b44afa754192933\": rpc error: code = NotFound desc = could not find container \"a2b5914fe7354791d6773f33ac9378c19a8e5b4afca7990b0b44afa754192933\": container with ID starting with a2b5914fe7354791d6773f33ac9378c19a8e5b4afca7990b0b44afa754192933 not found: ID does not exist" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.728828 4868 scope.go:117] "RemoveContainer" containerID="bbc47c78ecea10373d709f8016370bad805694e3d2800bf35760bd08eb36dc39" Dec 01 17:47:05 crc kubenswrapper[4868]: E1201 17:47:05.729145 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbc47c78ecea10373d709f8016370bad805694e3d2800bf35760bd08eb36dc39\": container with ID starting with bbc47c78ecea10373d709f8016370bad805694e3d2800bf35760bd08eb36dc39 not found: ID does not exist" containerID="bbc47c78ecea10373d709f8016370bad805694e3d2800bf35760bd08eb36dc39" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.729172 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbc47c78ecea10373d709f8016370bad805694e3d2800bf35760bd08eb36dc39"} err="failed to get container status \"bbc47c78ecea10373d709f8016370bad805694e3d2800bf35760bd08eb36dc39\": rpc error: code = NotFound desc = could not find container \"bbc47c78ecea10373d709f8016370bad805694e3d2800bf35760bd08eb36dc39\": container with ID starting with bbc47c78ecea10373d709f8016370bad805694e3d2800bf35760bd08eb36dc39 not found: ID does not exist" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.851443 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3ff4f488-06c8-4ceb-848a-4ba68b334752-pod-info\") pod \"3ff4f488-06c8-4ceb-848a-4ba68b334752\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.851492 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"3ff4f488-06c8-4ceb-848a-4ba68b334752\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.851519 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-confd\") pod \"3ff4f488-06c8-4ceb-848a-4ba68b334752\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.851567 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-tls\") pod \"3ff4f488-06c8-4ceb-848a-4ba68b334752\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.851596 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-erlang-cookie\") pod \"3ff4f488-06c8-4ceb-848a-4ba68b334752\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.851686 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgtg8\" (UniqueName: \"kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-kube-api-access-tgtg8\") pod \"3ff4f488-06c8-4ceb-848a-4ba68b334752\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.851705 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-plugins\") pod \"3ff4f488-06c8-4ceb-848a-4ba68b334752\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.851723 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-plugins-conf\") pod \"3ff4f488-06c8-4ceb-848a-4ba68b334752\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.851749 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3ff4f488-06c8-4ceb-848a-4ba68b334752-erlang-cookie-secret\") pod \"3ff4f488-06c8-4ceb-848a-4ba68b334752\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.851770 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-config-data\") pod \"3ff4f488-06c8-4ceb-848a-4ba68b334752\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.851805 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-server-conf\") pod \"3ff4f488-06c8-4ceb-848a-4ba68b334752\" (UID: \"3ff4f488-06c8-4ceb-848a-4ba68b334752\") " Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.853270 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "3ff4f488-06c8-4ceb-848a-4ba68b334752" (UID: "3ff4f488-06c8-4ceb-848a-4ba68b334752"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.853267 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "3ff4f488-06c8-4ceb-848a-4ba68b334752" (UID: "3ff4f488-06c8-4ceb-848a-4ba68b334752"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.853520 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "3ff4f488-06c8-4ceb-848a-4ba68b334752" (UID: "3ff4f488-06c8-4ceb-848a-4ba68b334752"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.858797 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ff4f488-06c8-4ceb-848a-4ba68b334752-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "3ff4f488-06c8-4ceb-848a-4ba68b334752" (UID: "3ff4f488-06c8-4ceb-848a-4ba68b334752"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.860441 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/3ff4f488-06c8-4ceb-848a-4ba68b334752-pod-info" (OuterVolumeSpecName: "pod-info") pod "3ff4f488-06c8-4ceb-848a-4ba68b334752" (UID: "3ff4f488-06c8-4ceb-848a-4ba68b334752"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.862424 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "3ff4f488-06c8-4ceb-848a-4ba68b334752" (UID: "3ff4f488-06c8-4ceb-848a-4ba68b334752"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.863408 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-kube-api-access-tgtg8" (OuterVolumeSpecName: "kube-api-access-tgtg8") pod "3ff4f488-06c8-4ceb-848a-4ba68b334752" (UID: "3ff4f488-06c8-4ceb-848a-4ba68b334752"). InnerVolumeSpecName "kube-api-access-tgtg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.863470 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "3ff4f488-06c8-4ceb-848a-4ba68b334752" (UID: "3ff4f488-06c8-4ceb-848a-4ba68b334752"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.866282 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.882802 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-config-data" (OuterVolumeSpecName: "config-data") pod "3ff4f488-06c8-4ceb-848a-4ba68b334752" (UID: "3ff4f488-06c8-4ceb-848a-4ba68b334752"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.918826 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-server-conf" (OuterVolumeSpecName: "server-conf") pod "3ff4f488-06c8-4ceb-848a-4ba68b334752" (UID: "3ff4f488-06c8-4ceb-848a-4ba68b334752"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.953444 4868 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3ff4f488-06c8-4ceb-848a-4ba68b334752-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.953767 4868 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.953778 4868 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.953788 4868 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.953817 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgtg8\" (UniqueName: \"kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-kube-api-access-tgtg8\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.953829 4868 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.953838 4868 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.953847 4868 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3ff4f488-06c8-4ceb-848a-4ba68b334752-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.953855 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.953863 4868 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3ff4f488-06c8-4ceb-848a-4ba68b334752-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.974214 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "3ff4f488-06c8-4ceb-848a-4ba68b334752" (UID: "3ff4f488-06c8-4ceb-848a-4ba68b334752"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:47:05 crc kubenswrapper[4868]: I1201 17:47:05.976992 4868 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.056012 4868 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.056056 4868 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3ff4f488-06c8-4ceb-848a-4ba68b334752-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.187902 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70aee7bf-d819-4355-a98c-4a106f3e6f16" path="/var/lib/kubelet/pods/70aee7bf-d819-4355-a98c-4a106f3e6f16/volumes" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.645494 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-f7kcn"] Dec 01 17:47:06 crc kubenswrapper[4868]: E1201 17:47:06.645919 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ff4f488-06c8-4ceb-848a-4ba68b334752" containerName="setup-container" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.645994 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ff4f488-06c8-4ceb-848a-4ba68b334752" containerName="setup-container" Dec 01 17:47:06 crc kubenswrapper[4868]: E1201 17:47:06.646043 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ff4f488-06c8-4ceb-848a-4ba68b334752" containerName="rabbitmq" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.646053 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ff4f488-06c8-4ceb-848a-4ba68b334752" containerName="rabbitmq" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.646266 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ff4f488-06c8-4ceb-848a-4ba68b334752" containerName="rabbitmq" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.647218 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.649215 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.669340 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-f7kcn"] Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.683308 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.698921 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b05cb910-a007-46d5-8182-eb1d30c5f953","Type":"ContainerStarted","Data":"9cf3101cb100c4b5cb4df5b7bd9932924dab60602c105850ae386e4d6c055bb2"} Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.738940 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.758167 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.767393 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-dns-svc\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.767447 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.767474 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vct9v\" (UniqueName: \"kubernetes.io/projected/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-kube-api-access-vct9v\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.768308 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.768437 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-config\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.768474 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.768554 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.774748 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.776705 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.779972 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.780189 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.780503 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.780694 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-v799m" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.780811 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.780922 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.781620 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.785707 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.870820 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.870930 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-config\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.870971 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.871007 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.871081 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-dns-svc\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.871107 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.871130 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vct9v\" (UniqueName: \"kubernetes.io/projected/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-kube-api-access-vct9v\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.872166 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-config\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.872209 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.872721 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.872820 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-dns-svc\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.873217 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.874620 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.975044 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.975173 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.975203 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.975232 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.975258 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnsxz\" (UniqueName: \"kubernetes.io/projected/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-kube-api-access-mnsxz\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.975284 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.975425 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.975571 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.975605 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.975686 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.975722 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:06 crc kubenswrapper[4868]: I1201 17:47:06.989734 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vct9v\" (UniqueName: \"kubernetes.io/projected/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-kube-api-access-vct9v\") pod \"dnsmasq-dns-d558885bc-f7kcn\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.077911 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.078034 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.078063 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.078088 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.078110 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnsxz\" (UniqueName: \"kubernetes.io/projected/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-kube-api-access-mnsxz\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.078139 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.078171 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.078219 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.078238 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.078271 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.078292 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.078586 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.078922 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.079572 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.079789 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.079998 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.080281 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.101640 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.102068 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnsxz\" (UniqueName: \"kubernetes.io/projected/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-kube-api-access-mnsxz\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.102475 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.104642 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.119111 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/12c9fc00-91a0-493e-b6b3-5d1a5ba7685c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.134502 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.267105 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.398165 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.711030 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b05cb910-a007-46d5-8182-eb1d30c5f953","Type":"ContainerStarted","Data":"d7eeda15a5a3a5591b252763ad434e859ed2cc81005a48e163bc37280e263e04"} Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.748429 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-f7kcn"] Dec 01 17:47:07 crc kubenswrapper[4868]: I1201 17:47:07.888071 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 17:47:07 crc kubenswrapper[4868]: W1201 17:47:07.899209 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12c9fc00_91a0_493e_b6b3_5d1a5ba7685c.slice/crio-208894fb40c5510342aaccf81c3e66cc0db8a4ed9c7010edcbbbbe98ad2e9a00 WatchSource:0}: Error finding container 208894fb40c5510342aaccf81c3e66cc0db8a4ed9c7010edcbbbbe98ad2e9a00: Status 404 returned error can't find the container with id 208894fb40c5510342aaccf81c3e66cc0db8a4ed9c7010edcbbbbe98ad2e9a00 Dec 01 17:47:08 crc kubenswrapper[4868]: I1201 17:47:08.182800 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ff4f488-06c8-4ceb-848a-4ba68b334752" path="/var/lib/kubelet/pods/3ff4f488-06c8-4ceb-848a-4ba68b334752/volumes" Dec 01 17:47:08 crc kubenswrapper[4868]: I1201 17:47:08.719761 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c","Type":"ContainerStarted","Data":"208894fb40c5510342aaccf81c3e66cc0db8a4ed9c7010edcbbbbe98ad2e9a00"} Dec 01 17:47:08 crc kubenswrapper[4868]: I1201 17:47:08.722031 4868 generic.go:334] "Generic (PLEG): container finished" podID="d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" containerID="4fed5588183c172d2de3f0d6274ed61ca0b55fcf5486935b687226aafe5f722b" exitCode=0 Dec 01 17:47:08 crc kubenswrapper[4868]: I1201 17:47:08.722128 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-f7kcn" event={"ID":"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2","Type":"ContainerDied","Data":"4fed5588183c172d2de3f0d6274ed61ca0b55fcf5486935b687226aafe5f722b"} Dec 01 17:47:08 crc kubenswrapper[4868]: I1201 17:47:08.722154 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-f7kcn" event={"ID":"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2","Type":"ContainerStarted","Data":"1833c37dfc3a6a2a1ec5ce45fd1f565c4ed7ef714c8386202826e4a74ede3389"} Dec 01 17:47:09 crc kubenswrapper[4868]: I1201 17:47:09.732726 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-f7kcn" event={"ID":"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2","Type":"ContainerStarted","Data":"c65acd3a3d8781f0647e5dc636838170352b1401a8a525b5a6e188f196e4560c"} Dec 01 17:47:09 crc kubenswrapper[4868]: I1201 17:47:09.732885 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:09 crc kubenswrapper[4868]: I1201 17:47:09.734877 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c","Type":"ContainerStarted","Data":"95584bc2e5dd7ab46abcf133007e7c596f04390bc48bd905543d82e53e8843ad"} Dec 01 17:47:09 crc kubenswrapper[4868]: I1201 17:47:09.751554 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-f7kcn" podStartSLOduration=3.751531288 podStartE2EDuration="3.751531288s" podCreationTimestamp="2025-12-01 17:47:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:47:09.750014549 +0000 UTC m=+1302.121124970" watchObservedRunningTime="2025-12-01 17:47:09.751531288 +0000 UTC m=+1302.122641699" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.268155 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.333889 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-pkhvw"] Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.335281 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" podUID="84e8b909-9eb9-4fdd-97fb-4b98256006aa" containerName="dnsmasq-dns" containerID="cri-o://093bbc91a298e5f48e40e82040e6a5b1956f89305df649897d67c8242fa90136" gracePeriod=10 Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.524007 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-4rnqg"] Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.526894 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.541073 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-4rnqg"] Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.610519 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.610707 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-config\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.610744 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.610789 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.610862 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zspl\" (UniqueName: \"kubernetes.io/projected/7480b802-459b-4029-9bc3-e3d7ba5099de-kube-api-access-4zspl\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.610884 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.610906 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.712896 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.712986 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.713067 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zspl\" (UniqueName: \"kubernetes.io/projected/7480b802-459b-4029-9bc3-e3d7ba5099de-kube-api-access-4zspl\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.713099 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.713134 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.713210 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.713310 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-config\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.715776 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-config\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.715795 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.716444 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.717962 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.720699 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.720757 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7480b802-459b-4029-9bc3-e3d7ba5099de-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.737711 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zspl\" (UniqueName: \"kubernetes.io/projected/7480b802-459b-4029-9bc3-e3d7ba5099de-kube-api-access-4zspl\") pod \"dnsmasq-dns-78c64bc9c5-4rnqg\" (UID: \"7480b802-459b-4029-9bc3-e3d7ba5099de\") " pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.813276 4868 generic.go:334] "Generic (PLEG): container finished" podID="84e8b909-9eb9-4fdd-97fb-4b98256006aa" containerID="093bbc91a298e5f48e40e82040e6a5b1956f89305df649897d67c8242fa90136" exitCode=0 Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.813324 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" event={"ID":"84e8b909-9eb9-4fdd-97fb-4b98256006aa","Type":"ContainerDied","Data":"093bbc91a298e5f48e40e82040e6a5b1956f89305df649897d67c8242fa90136"} Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.813350 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" event={"ID":"84e8b909-9eb9-4fdd-97fb-4b98256006aa","Type":"ContainerDied","Data":"0ad25d1c081d4456e72ca96552eed495b1c2d19eca88b82b43531660572cb982"} Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.813361 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ad25d1c081d4456e72ca96552eed495b1c2d19eca88b82b43531660572cb982" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.846920 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.868140 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.916860 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-dns-swift-storage-0\") pod \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.916919 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-ovsdbserver-nb\") pod \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.917072 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctkrj\" (UniqueName: \"kubernetes.io/projected/84e8b909-9eb9-4fdd-97fb-4b98256006aa-kube-api-access-ctkrj\") pod \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.917182 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-ovsdbserver-sb\") pod \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.917215 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-dns-svc\") pod \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.917403 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-config\") pod \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\" (UID: \"84e8b909-9eb9-4fdd-97fb-4b98256006aa\") " Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.926846 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84e8b909-9eb9-4fdd-97fb-4b98256006aa-kube-api-access-ctkrj" (OuterVolumeSpecName: "kube-api-access-ctkrj") pod "84e8b909-9eb9-4fdd-97fb-4b98256006aa" (UID: "84e8b909-9eb9-4fdd-97fb-4b98256006aa"). InnerVolumeSpecName "kube-api-access-ctkrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.987031 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-config" (OuterVolumeSpecName: "config") pod "84e8b909-9eb9-4fdd-97fb-4b98256006aa" (UID: "84e8b909-9eb9-4fdd-97fb-4b98256006aa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.987676 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "84e8b909-9eb9-4fdd-97fb-4b98256006aa" (UID: "84e8b909-9eb9-4fdd-97fb-4b98256006aa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.991167 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "84e8b909-9eb9-4fdd-97fb-4b98256006aa" (UID: "84e8b909-9eb9-4fdd-97fb-4b98256006aa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:17 crc kubenswrapper[4868]: I1201 17:47:17.997227 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "84e8b909-9eb9-4fdd-97fb-4b98256006aa" (UID: "84e8b909-9eb9-4fdd-97fb-4b98256006aa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:18 crc kubenswrapper[4868]: I1201 17:47:18.004596 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "84e8b909-9eb9-4fdd-97fb-4b98256006aa" (UID: "84e8b909-9eb9-4fdd-97fb-4b98256006aa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:18 crc kubenswrapper[4868]: I1201 17:47:18.020236 4868 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:18 crc kubenswrapper[4868]: I1201 17:47:18.020297 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:18 crc kubenswrapper[4868]: I1201 17:47:18.020308 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctkrj\" (UniqueName: \"kubernetes.io/projected/84e8b909-9eb9-4fdd-97fb-4b98256006aa-kube-api-access-ctkrj\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:18 crc kubenswrapper[4868]: I1201 17:47:18.020317 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:18 crc kubenswrapper[4868]: I1201 17:47:18.020326 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:18 crc kubenswrapper[4868]: I1201 17:47:18.020336 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84e8b909-9eb9-4fdd-97fb-4b98256006aa-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:18 crc kubenswrapper[4868]: I1201 17:47:18.397046 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-4rnqg"] Dec 01 17:47:18 crc kubenswrapper[4868]: I1201 17:47:18.822539 4868 generic.go:334] "Generic (PLEG): container finished" podID="7480b802-459b-4029-9bc3-e3d7ba5099de" containerID="d61f189841bbce5a543a823a111f1d4ef6a41e4e3b8adb6161fd1260cc3da02c" exitCode=0 Dec 01 17:47:18 crc kubenswrapper[4868]: I1201 17:47:18.822639 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" event={"ID":"7480b802-459b-4029-9bc3-e3d7ba5099de","Type":"ContainerDied","Data":"d61f189841bbce5a543a823a111f1d4ef6a41e4e3b8adb6161fd1260cc3da02c"} Dec 01 17:47:18 crc kubenswrapper[4868]: I1201 17:47:18.822964 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-pkhvw" Dec 01 17:47:18 crc kubenswrapper[4868]: I1201 17:47:18.823027 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" event={"ID":"7480b802-459b-4029-9bc3-e3d7ba5099de","Type":"ContainerStarted","Data":"751ecc12d6beee7106efe117ecfeae20ea63817f789a69654a249552138b723b"} Dec 01 17:47:18 crc kubenswrapper[4868]: I1201 17:47:18.871835 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-pkhvw"] Dec 01 17:47:18 crc kubenswrapper[4868]: I1201 17:47:18.880218 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-pkhvw"] Dec 01 17:47:19 crc kubenswrapper[4868]: I1201 17:47:19.833153 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" event={"ID":"7480b802-459b-4029-9bc3-e3d7ba5099de","Type":"ContainerStarted","Data":"a3373d3d9281f95f1f38446f41e3863d19e0477a6128cc3fb91e60e31f402b30"} Dec 01 17:47:19 crc kubenswrapper[4868]: I1201 17:47:19.833559 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:19 crc kubenswrapper[4868]: I1201 17:47:19.859567 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" podStartSLOduration=2.859549374 podStartE2EDuration="2.859549374s" podCreationTimestamp="2025-12-01 17:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:47:19.854010133 +0000 UTC m=+1312.225120544" watchObservedRunningTime="2025-12-01 17:47:19.859549374 +0000 UTC m=+1312.230659785" Dec 01 17:47:20 crc kubenswrapper[4868]: I1201 17:47:20.186838 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84e8b909-9eb9-4fdd-97fb-4b98256006aa" path="/var/lib/kubelet/pods/84e8b909-9eb9-4fdd-97fb-4b98256006aa/volumes" Dec 01 17:47:27 crc kubenswrapper[4868]: I1201 17:47:27.870262 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78c64bc9c5-4rnqg" Dec 01 17:47:27 crc kubenswrapper[4868]: I1201 17:47:27.988138 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-f7kcn"] Dec 01 17:47:27 crc kubenswrapper[4868]: I1201 17:47:27.988382 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-f7kcn" podUID="d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" containerName="dnsmasq-dns" containerID="cri-o://c65acd3a3d8781f0647e5dc636838170352b1401a8a525b5a6e188f196e4560c" gracePeriod=10 Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.494093 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.551097 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-ovsdbserver-sb\") pod \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.551188 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-dns-svc\") pod \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.551237 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vct9v\" (UniqueName: \"kubernetes.io/projected/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-kube-api-access-vct9v\") pod \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.551371 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-config\") pod \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.551435 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-openstack-edpm-ipam\") pod \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.551502 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-ovsdbserver-nb\") pod \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.551529 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-dns-swift-storage-0\") pod \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\" (UID: \"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2\") " Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.568667 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-kube-api-access-vct9v" (OuterVolumeSpecName: "kube-api-access-vct9v") pod "d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" (UID: "d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2"). InnerVolumeSpecName "kube-api-access-vct9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.618626 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" (UID: "d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.621088 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" (UID: "d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.630644 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" (UID: "d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.633048 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" (UID: "d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.637835 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" (UID: "d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.638862 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-config" (OuterVolumeSpecName: "config") pod "d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" (UID: "d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.654053 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.654085 4868 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.654099 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vct9v\" (UniqueName: \"kubernetes.io/projected/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-kube-api-access-vct9v\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.654114 4868 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-config\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.654125 4868 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.654136 4868 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.654146 4868 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.944608 4868 generic.go:334] "Generic (PLEG): container finished" podID="d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" containerID="c65acd3a3d8781f0647e5dc636838170352b1401a8a525b5a6e188f196e4560c" exitCode=0 Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.944665 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-f7kcn" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.944705 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-f7kcn" event={"ID":"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2","Type":"ContainerDied","Data":"c65acd3a3d8781f0647e5dc636838170352b1401a8a525b5a6e188f196e4560c"} Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.946220 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-f7kcn" event={"ID":"d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2","Type":"ContainerDied","Data":"1833c37dfc3a6a2a1ec5ce45fd1f565c4ed7ef714c8386202826e4a74ede3389"} Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.946252 4868 scope.go:117] "RemoveContainer" containerID="c65acd3a3d8781f0647e5dc636838170352b1401a8a525b5a6e188f196e4560c" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.979990 4868 scope.go:117] "RemoveContainer" containerID="4fed5588183c172d2de3f0d6274ed61ca0b55fcf5486935b687226aafe5f722b" Dec 01 17:47:28 crc kubenswrapper[4868]: I1201 17:47:28.991496 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-f7kcn"] Dec 01 17:47:29 crc kubenswrapper[4868]: I1201 17:47:29.012018 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-f7kcn"] Dec 01 17:47:29 crc kubenswrapper[4868]: I1201 17:47:29.021190 4868 scope.go:117] "RemoveContainer" containerID="c65acd3a3d8781f0647e5dc636838170352b1401a8a525b5a6e188f196e4560c" Dec 01 17:47:29 crc kubenswrapper[4868]: E1201 17:47:29.021652 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c65acd3a3d8781f0647e5dc636838170352b1401a8a525b5a6e188f196e4560c\": container with ID starting with c65acd3a3d8781f0647e5dc636838170352b1401a8a525b5a6e188f196e4560c not found: ID does not exist" containerID="c65acd3a3d8781f0647e5dc636838170352b1401a8a525b5a6e188f196e4560c" Dec 01 17:47:29 crc kubenswrapper[4868]: I1201 17:47:29.021719 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c65acd3a3d8781f0647e5dc636838170352b1401a8a525b5a6e188f196e4560c"} err="failed to get container status \"c65acd3a3d8781f0647e5dc636838170352b1401a8a525b5a6e188f196e4560c\": rpc error: code = NotFound desc = could not find container \"c65acd3a3d8781f0647e5dc636838170352b1401a8a525b5a6e188f196e4560c\": container with ID starting with c65acd3a3d8781f0647e5dc636838170352b1401a8a525b5a6e188f196e4560c not found: ID does not exist" Dec 01 17:47:29 crc kubenswrapper[4868]: I1201 17:47:29.021750 4868 scope.go:117] "RemoveContainer" containerID="4fed5588183c172d2de3f0d6274ed61ca0b55fcf5486935b687226aafe5f722b" Dec 01 17:47:29 crc kubenswrapper[4868]: E1201 17:47:29.022135 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fed5588183c172d2de3f0d6274ed61ca0b55fcf5486935b687226aafe5f722b\": container with ID starting with 4fed5588183c172d2de3f0d6274ed61ca0b55fcf5486935b687226aafe5f722b not found: ID does not exist" containerID="4fed5588183c172d2de3f0d6274ed61ca0b55fcf5486935b687226aafe5f722b" Dec 01 17:47:29 crc kubenswrapper[4868]: I1201 17:47:29.022169 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fed5588183c172d2de3f0d6274ed61ca0b55fcf5486935b687226aafe5f722b"} err="failed to get container status \"4fed5588183c172d2de3f0d6274ed61ca0b55fcf5486935b687226aafe5f722b\": rpc error: code = NotFound desc = could not find container \"4fed5588183c172d2de3f0d6274ed61ca0b55fcf5486935b687226aafe5f722b\": container with ID starting with 4fed5588183c172d2de3f0d6274ed61ca0b55fcf5486935b687226aafe5f722b not found: ID does not exist" Dec 01 17:47:30 crc kubenswrapper[4868]: I1201 17:47:30.193382 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" path="/var/lib/kubelet/pods/d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2/volumes" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.537730 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb"] Dec 01 17:47:36 crc kubenswrapper[4868]: E1201 17:47:36.538721 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" containerName="dnsmasq-dns" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.538739 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" containerName="dnsmasq-dns" Dec 01 17:47:36 crc kubenswrapper[4868]: E1201 17:47:36.538771 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e8b909-9eb9-4fdd-97fb-4b98256006aa" containerName="dnsmasq-dns" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.538779 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e8b909-9eb9-4fdd-97fb-4b98256006aa" containerName="dnsmasq-dns" Dec 01 17:47:36 crc kubenswrapper[4868]: E1201 17:47:36.538795 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" containerName="init" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.538805 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" containerName="init" Dec 01 17:47:36 crc kubenswrapper[4868]: E1201 17:47:36.538814 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e8b909-9eb9-4fdd-97fb-4b98256006aa" containerName="init" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.538821 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e8b909-9eb9-4fdd-97fb-4b98256006aa" containerName="init" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.539045 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="84e8b909-9eb9-4fdd-97fb-4b98256006aa" containerName="dnsmasq-dns" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.539062 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="d15d99ed-2a81-4cf8-b84d-cc72cda3f9e2" containerName="dnsmasq-dns" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.539804 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.542474 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.542598 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.544400 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.545104 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.551388 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb"] Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.610204 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.610260 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwkb2\" (UniqueName: \"kubernetes.io/projected/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-kube-api-access-bwkb2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.610307 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.610357 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.712463 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.712579 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.712615 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwkb2\" (UniqueName: \"kubernetes.io/projected/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-kube-api-access-bwkb2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.712658 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.718526 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.718810 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.724933 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.728735 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwkb2\" (UniqueName: \"kubernetes.io/projected/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-kube-api-access-bwkb2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:47:36 crc kubenswrapper[4868]: I1201 17:47:36.874013 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:47:37 crc kubenswrapper[4868]: I1201 17:47:37.420202 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb"] Dec 01 17:47:37 crc kubenswrapper[4868]: W1201 17:47:37.428767 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf36695d5_3d20_4e3a_9ef3_63cc50a3445c.slice/crio-cbf78c1679f0fe31119ce3f0cc60553a16d4f92f654f7e33457dcb949f9064f7 WatchSource:0}: Error finding container cbf78c1679f0fe31119ce3f0cc60553a16d4f92f654f7e33457dcb949f9064f7: Status 404 returned error can't find the container with id cbf78c1679f0fe31119ce3f0cc60553a16d4f92f654f7e33457dcb949f9064f7 Dec 01 17:47:38 crc kubenswrapper[4868]: I1201 17:47:38.047003 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" event={"ID":"f36695d5-3d20-4e3a-9ef3-63cc50a3445c","Type":"ContainerStarted","Data":"cbf78c1679f0fe31119ce3f0cc60553a16d4f92f654f7e33457dcb949f9064f7"} Dec 01 17:47:40 crc kubenswrapper[4868]: I1201 17:47:40.070233 4868 generic.go:334] "Generic (PLEG): container finished" podID="b05cb910-a007-46d5-8182-eb1d30c5f953" containerID="d7eeda15a5a3a5591b252763ad434e859ed2cc81005a48e163bc37280e263e04" exitCode=0 Dec 01 17:47:40 crc kubenswrapper[4868]: I1201 17:47:40.070327 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b05cb910-a007-46d5-8182-eb1d30c5f953","Type":"ContainerDied","Data":"d7eeda15a5a3a5591b252763ad434e859ed2cc81005a48e163bc37280e263e04"} Dec 01 17:47:41 crc kubenswrapper[4868]: I1201 17:47:41.095624 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b05cb910-a007-46d5-8182-eb1d30c5f953","Type":"ContainerStarted","Data":"4479638ab2275edcd84dd7dd59c8b4c8eaee83372d768577eb0b10456872b40a"} Dec 01 17:47:41 crc kubenswrapper[4868]: I1201 17:47:41.096639 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 01 17:47:41 crc kubenswrapper[4868]: I1201 17:47:41.136119 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.136093899 podStartE2EDuration="37.136093899s" podCreationTimestamp="2025-12-01 17:47:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:47:41.130443316 +0000 UTC m=+1333.501553727" watchObservedRunningTime="2025-12-01 17:47:41.136093899 +0000 UTC m=+1333.507204310" Dec 01 17:47:42 crc kubenswrapper[4868]: I1201 17:47:42.110673 4868 generic.go:334] "Generic (PLEG): container finished" podID="12c9fc00-91a0-493e-b6b3-5d1a5ba7685c" containerID="95584bc2e5dd7ab46abcf133007e7c596f04390bc48bd905543d82e53e8843ad" exitCode=0 Dec 01 17:47:42 crc kubenswrapper[4868]: I1201 17:47:42.111126 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c","Type":"ContainerDied","Data":"95584bc2e5dd7ab46abcf133007e7c596f04390bc48bd905543d82e53e8843ad"} Dec 01 17:47:47 crc kubenswrapper[4868]: I1201 17:47:47.164763 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"12c9fc00-91a0-493e-b6b3-5d1a5ba7685c","Type":"ContainerStarted","Data":"0c9e6af6dfb29f73d5408144a42b3491be42fe3ec31e93e6b093384f4647aa26"} Dec 01 17:47:47 crc kubenswrapper[4868]: I1201 17:47:47.165509 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:47 crc kubenswrapper[4868]: I1201 17:47:47.165912 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" event={"ID":"f36695d5-3d20-4e3a-9ef3-63cc50a3445c","Type":"ContainerStarted","Data":"008375ad16a01abfabffb08f36e162220a7224d8d7f6f20ec53b6e1cf0e2c8b9"} Dec 01 17:47:47 crc kubenswrapper[4868]: I1201 17:47:47.189224 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=41.189203268 podStartE2EDuration="41.189203268s" podCreationTimestamp="2025-12-01 17:47:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:47:47.186798296 +0000 UTC m=+1339.557908707" watchObservedRunningTime="2025-12-01 17:47:47.189203268 +0000 UTC m=+1339.560313679" Dec 01 17:47:47 crc kubenswrapper[4868]: I1201 17:47:47.207183 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" podStartSLOduration=1.8936396009999998 podStartE2EDuration="11.207165593s" podCreationTimestamp="2025-12-01 17:47:36 +0000 UTC" firstStartedPulling="2025-12-01 17:47:37.434481048 +0000 UTC m=+1329.805591459" lastFinishedPulling="2025-12-01 17:47:46.74800705 +0000 UTC m=+1339.119117451" observedRunningTime="2025-12-01 17:47:47.203743256 +0000 UTC m=+1339.574853667" watchObservedRunningTime="2025-12-01 17:47:47.207165593 +0000 UTC m=+1339.578275994" Dec 01 17:47:55 crc kubenswrapper[4868]: I1201 17:47:55.346243 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 01 17:47:57 crc kubenswrapper[4868]: I1201 17:47:57.401112 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 01 17:47:59 crc kubenswrapper[4868]: I1201 17:47:59.279146 4868 generic.go:334] "Generic (PLEG): container finished" podID="f36695d5-3d20-4e3a-9ef3-63cc50a3445c" containerID="008375ad16a01abfabffb08f36e162220a7224d8d7f6f20ec53b6e1cf0e2c8b9" exitCode=0 Dec 01 17:47:59 crc kubenswrapper[4868]: I1201 17:47:59.279250 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" event={"ID":"f36695d5-3d20-4e3a-9ef3-63cc50a3445c","Type":"ContainerDied","Data":"008375ad16a01abfabffb08f36e162220a7224d8d7f6f20ec53b6e1cf0e2c8b9"} Dec 01 17:48:00 crc kubenswrapper[4868]: I1201 17:48:00.722160 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:48:00 crc kubenswrapper[4868]: I1201 17:48:00.833771 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-ssh-key\") pod \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " Dec 01 17:48:00 crc kubenswrapper[4868]: I1201 17:48:00.833913 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-repo-setup-combined-ca-bundle\") pod \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " Dec 01 17:48:00 crc kubenswrapper[4868]: I1201 17:48:00.834134 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwkb2\" (UniqueName: \"kubernetes.io/projected/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-kube-api-access-bwkb2\") pod \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " Dec 01 17:48:00 crc kubenswrapper[4868]: I1201 17:48:00.834434 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-inventory\") pod \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\" (UID: \"f36695d5-3d20-4e3a-9ef3-63cc50a3445c\") " Dec 01 17:48:00 crc kubenswrapper[4868]: I1201 17:48:00.841285 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-kube-api-access-bwkb2" (OuterVolumeSpecName: "kube-api-access-bwkb2") pod "f36695d5-3d20-4e3a-9ef3-63cc50a3445c" (UID: "f36695d5-3d20-4e3a-9ef3-63cc50a3445c"). InnerVolumeSpecName "kube-api-access-bwkb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:48:00 crc kubenswrapper[4868]: I1201 17:48:00.847341 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "f36695d5-3d20-4e3a-9ef3-63cc50a3445c" (UID: "f36695d5-3d20-4e3a-9ef3-63cc50a3445c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:48:00 crc kubenswrapper[4868]: I1201 17:48:00.866693 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f36695d5-3d20-4e3a-9ef3-63cc50a3445c" (UID: "f36695d5-3d20-4e3a-9ef3-63cc50a3445c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:48:00 crc kubenswrapper[4868]: I1201 17:48:00.873935 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-inventory" (OuterVolumeSpecName: "inventory") pod "f36695d5-3d20-4e3a-9ef3-63cc50a3445c" (UID: "f36695d5-3d20-4e3a-9ef3-63cc50a3445c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:48:00 crc kubenswrapper[4868]: I1201 17:48:00.936926 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwkb2\" (UniqueName: \"kubernetes.io/projected/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-kube-api-access-bwkb2\") on node \"crc\" DevicePath \"\"" Dec 01 17:48:00 crc kubenswrapper[4868]: I1201 17:48:00.936971 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:48:00 crc kubenswrapper[4868]: I1201 17:48:00.936982 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:48:00 crc kubenswrapper[4868]: I1201 17:48:00.936992 4868 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f36695d5-3d20-4e3a-9ef3-63cc50a3445c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.295967 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" event={"ID":"f36695d5-3d20-4e3a-9ef3-63cc50a3445c","Type":"ContainerDied","Data":"cbf78c1679f0fe31119ce3f0cc60553a16d4f92f654f7e33457dcb949f9064f7"} Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.296021 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbf78c1679f0fe31119ce3f0cc60553a16d4f92f654f7e33457dcb949f9064f7" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.296086 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.372804 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm"] Dec 01 17:48:01 crc kubenswrapper[4868]: E1201 17:48:01.374237 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f36695d5-3d20-4e3a-9ef3-63cc50a3445c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.374270 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f36695d5-3d20-4e3a-9ef3-63cc50a3445c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.374528 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f36695d5-3d20-4e3a-9ef3-63cc50a3445c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.375348 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.376891 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.377385 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.378518 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.381425 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.388543 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm"] Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.446793 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e989defc-ac68-4417-9bf3-f5962bfdb601-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vvbkm\" (UID: \"e989defc-ac68-4417-9bf3-f5962bfdb601\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.447200 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lcc5\" (UniqueName: \"kubernetes.io/projected/e989defc-ac68-4417-9bf3-f5962bfdb601-kube-api-access-4lcc5\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vvbkm\" (UID: \"e989defc-ac68-4417-9bf3-f5962bfdb601\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.447286 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e989defc-ac68-4417-9bf3-f5962bfdb601-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vvbkm\" (UID: \"e989defc-ac68-4417-9bf3-f5962bfdb601\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.549036 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e989defc-ac68-4417-9bf3-f5962bfdb601-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vvbkm\" (UID: \"e989defc-ac68-4417-9bf3-f5962bfdb601\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.549120 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e989defc-ac68-4417-9bf3-f5962bfdb601-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vvbkm\" (UID: \"e989defc-ac68-4417-9bf3-f5962bfdb601\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.549182 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lcc5\" (UniqueName: \"kubernetes.io/projected/e989defc-ac68-4417-9bf3-f5962bfdb601-kube-api-access-4lcc5\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vvbkm\" (UID: \"e989defc-ac68-4417-9bf3-f5962bfdb601\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.556521 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e989defc-ac68-4417-9bf3-f5962bfdb601-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vvbkm\" (UID: \"e989defc-ac68-4417-9bf3-f5962bfdb601\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.556522 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e989defc-ac68-4417-9bf3-f5962bfdb601-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vvbkm\" (UID: \"e989defc-ac68-4417-9bf3-f5962bfdb601\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.565333 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lcc5\" (UniqueName: \"kubernetes.io/projected/e989defc-ac68-4417-9bf3-f5962bfdb601-kube-api-access-4lcc5\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vvbkm\" (UID: \"e989defc-ac68-4417-9bf3-f5962bfdb601\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" Dec 01 17:48:01 crc kubenswrapper[4868]: I1201 17:48:01.694530 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" Dec 01 17:48:02 crc kubenswrapper[4868]: I1201 17:48:02.223714 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm"] Dec 01 17:48:02 crc kubenswrapper[4868]: W1201 17:48:02.233172 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode989defc_ac68_4417_9bf3_f5962bfdb601.slice/crio-ce9a8cb7c0e0e409bdf04e6fdda2c8b68c9ea078480e42aa189e655099288e90 WatchSource:0}: Error finding container ce9a8cb7c0e0e409bdf04e6fdda2c8b68c9ea078480e42aa189e655099288e90: Status 404 returned error can't find the container with id ce9a8cb7c0e0e409bdf04e6fdda2c8b68c9ea078480e42aa189e655099288e90 Dec 01 17:48:02 crc kubenswrapper[4868]: I1201 17:48:02.305251 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" event={"ID":"e989defc-ac68-4417-9bf3-f5962bfdb601","Type":"ContainerStarted","Data":"ce9a8cb7c0e0e409bdf04e6fdda2c8b68c9ea078480e42aa189e655099288e90"} Dec 01 17:48:04 crc kubenswrapper[4868]: I1201 17:48:04.338390 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" event={"ID":"e989defc-ac68-4417-9bf3-f5962bfdb601","Type":"ContainerStarted","Data":"c638866daa20d8072b9a42b179f7af0dbdcd49047d4697be842720d4afd9e8b9"} Dec 01 17:48:04 crc kubenswrapper[4868]: I1201 17:48:04.376258 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" podStartSLOduration=2.5146745470000003 podStartE2EDuration="3.376235274s" podCreationTimestamp="2025-12-01 17:48:01 +0000 UTC" firstStartedPulling="2025-12-01 17:48:02.236792504 +0000 UTC m=+1354.607902915" lastFinishedPulling="2025-12-01 17:48:03.098353231 +0000 UTC m=+1355.469463642" observedRunningTime="2025-12-01 17:48:04.350387849 +0000 UTC m=+1356.721498260" watchObservedRunningTime="2025-12-01 17:48:04.376235274 +0000 UTC m=+1356.747345685" Dec 01 17:48:06 crc kubenswrapper[4868]: I1201 17:48:06.356809 4868 generic.go:334] "Generic (PLEG): container finished" podID="e989defc-ac68-4417-9bf3-f5962bfdb601" containerID="c638866daa20d8072b9a42b179f7af0dbdcd49047d4697be842720d4afd9e8b9" exitCode=0 Dec 01 17:48:06 crc kubenswrapper[4868]: I1201 17:48:06.356899 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" event={"ID":"e989defc-ac68-4417-9bf3-f5962bfdb601","Type":"ContainerDied","Data":"c638866daa20d8072b9a42b179f7af0dbdcd49047d4697be842720d4afd9e8b9"} Dec 01 17:48:07 crc kubenswrapper[4868]: I1201 17:48:07.779400 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" Dec 01 17:48:07 crc kubenswrapper[4868]: I1201 17:48:07.870144 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lcc5\" (UniqueName: \"kubernetes.io/projected/e989defc-ac68-4417-9bf3-f5962bfdb601-kube-api-access-4lcc5\") pod \"e989defc-ac68-4417-9bf3-f5962bfdb601\" (UID: \"e989defc-ac68-4417-9bf3-f5962bfdb601\") " Dec 01 17:48:07 crc kubenswrapper[4868]: I1201 17:48:07.870211 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e989defc-ac68-4417-9bf3-f5962bfdb601-inventory\") pod \"e989defc-ac68-4417-9bf3-f5962bfdb601\" (UID: \"e989defc-ac68-4417-9bf3-f5962bfdb601\") " Dec 01 17:48:07 crc kubenswrapper[4868]: I1201 17:48:07.870387 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e989defc-ac68-4417-9bf3-f5962bfdb601-ssh-key\") pod \"e989defc-ac68-4417-9bf3-f5962bfdb601\" (UID: \"e989defc-ac68-4417-9bf3-f5962bfdb601\") " Dec 01 17:48:07 crc kubenswrapper[4868]: I1201 17:48:07.878240 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e989defc-ac68-4417-9bf3-f5962bfdb601-kube-api-access-4lcc5" (OuterVolumeSpecName: "kube-api-access-4lcc5") pod "e989defc-ac68-4417-9bf3-f5962bfdb601" (UID: "e989defc-ac68-4417-9bf3-f5962bfdb601"). InnerVolumeSpecName "kube-api-access-4lcc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:48:07 crc kubenswrapper[4868]: I1201 17:48:07.905571 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e989defc-ac68-4417-9bf3-f5962bfdb601-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e989defc-ac68-4417-9bf3-f5962bfdb601" (UID: "e989defc-ac68-4417-9bf3-f5962bfdb601"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:48:07 crc kubenswrapper[4868]: I1201 17:48:07.916694 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e989defc-ac68-4417-9bf3-f5962bfdb601-inventory" (OuterVolumeSpecName: "inventory") pod "e989defc-ac68-4417-9bf3-f5962bfdb601" (UID: "e989defc-ac68-4417-9bf3-f5962bfdb601"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:48:07 crc kubenswrapper[4868]: I1201 17:48:07.972403 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e989defc-ac68-4417-9bf3-f5962bfdb601-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:48:07 crc kubenswrapper[4868]: I1201 17:48:07.972439 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lcc5\" (UniqueName: \"kubernetes.io/projected/e989defc-ac68-4417-9bf3-f5962bfdb601-kube-api-access-4lcc5\") on node \"crc\" DevicePath \"\"" Dec 01 17:48:07 crc kubenswrapper[4868]: I1201 17:48:07.972451 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e989defc-ac68-4417-9bf3-f5962bfdb601-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.375454 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" event={"ID":"e989defc-ac68-4417-9bf3-f5962bfdb601","Type":"ContainerDied","Data":"ce9a8cb7c0e0e409bdf04e6fdda2c8b68c9ea078480e42aa189e655099288e90"} Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.375500 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce9a8cb7c0e0e409bdf04e6fdda2c8b68c9ea078480e42aa189e655099288e90" Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.375537 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vvbkm" Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.520423 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp"] Dec 01 17:48:08 crc kubenswrapper[4868]: E1201 17:48:08.520984 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e989defc-ac68-4417-9bf3-f5962bfdb601" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.521010 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="e989defc-ac68-4417-9bf3-f5962bfdb601" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.521271 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="e989defc-ac68-4417-9bf3-f5962bfdb601" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.521929 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.527753 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.527792 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.533101 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp"] Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.543025 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.543725 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.591881 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.592163 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.592232 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:48:08 crc kubenswrapper[4868]: I1201 17:48:08.592261 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdq2j\" (UniqueName: \"kubernetes.io/projected/af68e45f-ae90-4a6b-a189-7a97b42d0b95-kube-api-access-xdq2j\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:48:09 crc kubenswrapper[4868]: I1201 17:48:08.694308 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:48:09 crc kubenswrapper[4868]: I1201 17:48:08.694432 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:48:09 crc kubenswrapper[4868]: I1201 17:48:08.694461 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:48:09 crc kubenswrapper[4868]: I1201 17:48:08.694483 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdq2j\" (UniqueName: \"kubernetes.io/projected/af68e45f-ae90-4a6b-a189-7a97b42d0b95-kube-api-access-xdq2j\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:48:09 crc kubenswrapper[4868]: I1201 17:48:08.698619 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:48:09 crc kubenswrapper[4868]: I1201 17:48:08.699117 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:48:09 crc kubenswrapper[4868]: I1201 17:48:08.702703 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:48:09 crc kubenswrapper[4868]: I1201 17:48:08.711613 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdq2j\" (UniqueName: \"kubernetes.io/projected/af68e45f-ae90-4a6b-a189-7a97b42d0b95-kube-api-access-xdq2j\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:48:09 crc kubenswrapper[4868]: I1201 17:48:08.844652 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:48:09 crc kubenswrapper[4868]: I1201 17:48:09.730093 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp"] Dec 01 17:48:10 crc kubenswrapper[4868]: I1201 17:48:10.397402 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" event={"ID":"af68e45f-ae90-4a6b-a189-7a97b42d0b95","Type":"ContainerStarted","Data":"a916122ca7b4daa94a39a28317643c51842241ceabd9b123b9c5b6c9ebc13430"} Dec 01 17:48:11 crc kubenswrapper[4868]: I1201 17:48:11.423752 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" event={"ID":"af68e45f-ae90-4a6b-a189-7a97b42d0b95","Type":"ContainerStarted","Data":"b0f54559e1315b65f72c499bf86eaf341d919617b878f84b87771d6e3ed337b1"} Dec 01 17:48:11 crc kubenswrapper[4868]: I1201 17:48:11.447326 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" podStartSLOduration=2.904699982 podStartE2EDuration="3.447302189s" podCreationTimestamp="2025-12-01 17:48:08 +0000 UTC" firstStartedPulling="2025-12-01 17:48:09.724266578 +0000 UTC m=+1362.095376989" lastFinishedPulling="2025-12-01 17:48:10.266868785 +0000 UTC m=+1362.637979196" observedRunningTime="2025-12-01 17:48:11.443308801 +0000 UTC m=+1363.814419232" watchObservedRunningTime="2025-12-01 17:48:11.447302189 +0000 UTC m=+1363.818412610" Dec 01 17:48:38 crc kubenswrapper[4868]: I1201 17:48:38.774247 4868 scope.go:117] "RemoveContainer" containerID="4707587d3ff16d0ec42fd9af5d40088384976db5daa980cb2f8997c715f84231" Dec 01 17:48:55 crc kubenswrapper[4868]: I1201 17:48:55.904470 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:48:55 crc kubenswrapper[4868]: I1201 17:48:55.906023 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:49:25 crc kubenswrapper[4868]: I1201 17:49:25.904810 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:49:25 crc kubenswrapper[4868]: I1201 17:49:25.905538 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:49:38 crc kubenswrapper[4868]: I1201 17:49:38.860307 4868 scope.go:117] "RemoveContainer" containerID="345f997cae8a892e373c6b5bcfb2d6321751e959e76a526501d682d8bbac14fb" Dec 01 17:49:55 crc kubenswrapper[4868]: I1201 17:49:55.904366 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:49:55 crc kubenswrapper[4868]: I1201 17:49:55.904915 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:49:55 crc kubenswrapper[4868]: I1201 17:49:55.904985 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:49:55 crc kubenswrapper[4868]: I1201 17:49:55.905790 4868 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"998e9f9a5d44e576ae1e608687cc6cddbcaf1995065246b296eae79bc764efdb"} pod="openshift-machine-config-operator/machine-config-daemon-njgpn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 17:49:55 crc kubenswrapper[4868]: I1201 17:49:55.905859 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" containerID="cri-o://998e9f9a5d44e576ae1e608687cc6cddbcaf1995065246b296eae79bc764efdb" gracePeriod=600 Dec 01 17:49:56 crc kubenswrapper[4868]: I1201 17:49:56.483650 4868 generic.go:334] "Generic (PLEG): container finished" podID="ec4e08bf-df56-4490-8e79-a4175c727405" containerID="998e9f9a5d44e576ae1e608687cc6cddbcaf1995065246b296eae79bc764efdb" exitCode=0 Dec 01 17:49:56 crc kubenswrapper[4868]: I1201 17:49:56.483727 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerDied","Data":"998e9f9a5d44e576ae1e608687cc6cddbcaf1995065246b296eae79bc764efdb"} Dec 01 17:49:56 crc kubenswrapper[4868]: I1201 17:49:56.484097 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded"} Dec 01 17:49:56 crc kubenswrapper[4868]: I1201 17:49:56.484122 4868 scope.go:117] "RemoveContainer" containerID="35d0f81eb6527dda2c61338bf9c12106c16489b5c045a71ea474df12051c7650" Dec 01 17:50:55 crc kubenswrapper[4868]: I1201 17:50:55.639664 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qdlmm"] Dec 01 17:50:55 crc kubenswrapper[4868]: I1201 17:50:55.652551 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:50:55 crc kubenswrapper[4868]: I1201 17:50:55.657070 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qdlmm"] Dec 01 17:50:55 crc kubenswrapper[4868]: I1201 17:50:55.702711 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-catalog-content\") pod \"community-operators-qdlmm\" (UID: \"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f\") " pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:50:55 crc kubenswrapper[4868]: I1201 17:50:55.702810 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8fsf\" (UniqueName: \"kubernetes.io/projected/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-kube-api-access-h8fsf\") pod \"community-operators-qdlmm\" (UID: \"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f\") " pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:50:55 crc kubenswrapper[4868]: I1201 17:50:55.702972 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-utilities\") pod \"community-operators-qdlmm\" (UID: \"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f\") " pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:50:55 crc kubenswrapper[4868]: I1201 17:50:55.804450 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8fsf\" (UniqueName: \"kubernetes.io/projected/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-kube-api-access-h8fsf\") pod \"community-operators-qdlmm\" (UID: \"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f\") " pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:50:55 crc kubenswrapper[4868]: I1201 17:50:55.804616 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-utilities\") pod \"community-operators-qdlmm\" (UID: \"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f\") " pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:50:55 crc kubenswrapper[4868]: I1201 17:50:55.804675 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-catalog-content\") pod \"community-operators-qdlmm\" (UID: \"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f\") " pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:50:55 crc kubenswrapper[4868]: I1201 17:50:55.805265 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-utilities\") pod \"community-operators-qdlmm\" (UID: \"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f\") " pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:50:55 crc kubenswrapper[4868]: I1201 17:50:55.805286 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-catalog-content\") pod \"community-operators-qdlmm\" (UID: \"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f\") " pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:50:55 crc kubenswrapper[4868]: I1201 17:50:55.823906 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8fsf\" (UniqueName: \"kubernetes.io/projected/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-kube-api-access-h8fsf\") pod \"community-operators-qdlmm\" (UID: \"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f\") " pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:50:55 crc kubenswrapper[4868]: I1201 17:50:55.988869 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:50:56 crc kubenswrapper[4868]: I1201 17:50:56.523903 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qdlmm"] Dec 01 17:50:57 crc kubenswrapper[4868]: I1201 17:50:57.100446 4868 generic.go:334] "Generic (PLEG): container finished" podID="4fe4c201-9b3b-4cfe-a020-48a47bac2e9f" containerID="b00eb3ceaa7834244a3a0d2d81cb06509d7eee8d5a8670af12932cb73d9fc00f" exitCode=0 Dec 01 17:50:57 crc kubenswrapper[4868]: I1201 17:50:57.100510 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qdlmm" event={"ID":"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f","Type":"ContainerDied","Data":"b00eb3ceaa7834244a3a0d2d81cb06509d7eee8d5a8670af12932cb73d9fc00f"} Dec 01 17:50:57 crc kubenswrapper[4868]: I1201 17:50:57.100733 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qdlmm" event={"ID":"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f","Type":"ContainerStarted","Data":"1beba41baf7abdb5452ca936bc050043e6448320b8a9bd8f58bcc6fc3ac049e3"} Dec 01 17:50:58 crc kubenswrapper[4868]: I1201 17:50:58.115063 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qdlmm" event={"ID":"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f","Type":"ContainerStarted","Data":"641f76135e1595e6e842926fa8df13749a638d0ab7b38351907d56b3ad3c3596"} Dec 01 17:50:59 crc kubenswrapper[4868]: I1201 17:50:59.125241 4868 generic.go:334] "Generic (PLEG): container finished" podID="4fe4c201-9b3b-4cfe-a020-48a47bac2e9f" containerID="641f76135e1595e6e842926fa8df13749a638d0ab7b38351907d56b3ad3c3596" exitCode=0 Dec 01 17:50:59 crc kubenswrapper[4868]: I1201 17:50:59.125336 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qdlmm" event={"ID":"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f","Type":"ContainerDied","Data":"641f76135e1595e6e842926fa8df13749a638d0ab7b38351907d56b3ad3c3596"} Dec 01 17:51:00 crc kubenswrapper[4868]: I1201 17:51:00.138434 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qdlmm" event={"ID":"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f","Type":"ContainerStarted","Data":"1d371a066df984ec7e71cb33824e8997f0417a05a70e89e69f04d0a5d1eacb13"} Dec 01 17:51:00 crc kubenswrapper[4868]: I1201 17:51:00.168594 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qdlmm" podStartSLOduration=2.5034667600000002 podStartE2EDuration="5.168575518s" podCreationTimestamp="2025-12-01 17:50:55 +0000 UTC" firstStartedPulling="2025-12-01 17:50:57.102768755 +0000 UTC m=+1529.473879166" lastFinishedPulling="2025-12-01 17:50:59.767877513 +0000 UTC m=+1532.138987924" observedRunningTime="2025-12-01 17:51:00.161520297 +0000 UTC m=+1532.532630718" watchObservedRunningTime="2025-12-01 17:51:00.168575518 +0000 UTC m=+1532.539685929" Dec 01 17:51:05 crc kubenswrapper[4868]: I1201 17:51:05.989908 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:51:05 crc kubenswrapper[4868]: I1201 17:51:05.991469 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:51:06 crc kubenswrapper[4868]: I1201 17:51:06.046684 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:51:06 crc kubenswrapper[4868]: I1201 17:51:06.231725 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:51:06 crc kubenswrapper[4868]: I1201 17:51:06.280651 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qdlmm"] Dec 01 17:51:08 crc kubenswrapper[4868]: I1201 17:51:08.220666 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qdlmm" podUID="4fe4c201-9b3b-4cfe-a020-48a47bac2e9f" containerName="registry-server" containerID="cri-o://1d371a066df984ec7e71cb33824e8997f0417a05a70e89e69f04d0a5d1eacb13" gracePeriod=2 Dec 01 17:51:08 crc kubenswrapper[4868]: I1201 17:51:08.716262 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:51:08 crc kubenswrapper[4868]: I1201 17:51:08.760090 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-utilities\") pod \"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f\" (UID: \"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f\") " Dec 01 17:51:08 crc kubenswrapper[4868]: I1201 17:51:08.760242 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8fsf\" (UniqueName: \"kubernetes.io/projected/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-kube-api-access-h8fsf\") pod \"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f\" (UID: \"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f\") " Dec 01 17:51:08 crc kubenswrapper[4868]: I1201 17:51:08.760277 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-catalog-content\") pod \"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f\" (UID: \"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f\") " Dec 01 17:51:08 crc kubenswrapper[4868]: I1201 17:51:08.764743 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-utilities" (OuterVolumeSpecName: "utilities") pod "4fe4c201-9b3b-4cfe-a020-48a47bac2e9f" (UID: "4fe4c201-9b3b-4cfe-a020-48a47bac2e9f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:51:08 crc kubenswrapper[4868]: I1201 17:51:08.775304 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-kube-api-access-h8fsf" (OuterVolumeSpecName: "kube-api-access-h8fsf") pod "4fe4c201-9b3b-4cfe-a020-48a47bac2e9f" (UID: "4fe4c201-9b3b-4cfe-a020-48a47bac2e9f"). InnerVolumeSpecName "kube-api-access-h8fsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:51:08 crc kubenswrapper[4868]: I1201 17:51:08.809312 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4fe4c201-9b3b-4cfe-a020-48a47bac2e9f" (UID: "4fe4c201-9b3b-4cfe-a020-48a47bac2e9f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:51:08 crc kubenswrapper[4868]: I1201 17:51:08.863107 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:51:08 crc kubenswrapper[4868]: I1201 17:51:08.863208 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8fsf\" (UniqueName: \"kubernetes.io/projected/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-kube-api-access-h8fsf\") on node \"crc\" DevicePath \"\"" Dec 01 17:51:08 crc kubenswrapper[4868]: I1201 17:51:08.863221 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:51:09 crc kubenswrapper[4868]: I1201 17:51:09.232220 4868 generic.go:334] "Generic (PLEG): container finished" podID="4fe4c201-9b3b-4cfe-a020-48a47bac2e9f" containerID="1d371a066df984ec7e71cb33824e8997f0417a05a70e89e69f04d0a5d1eacb13" exitCode=0 Dec 01 17:51:09 crc kubenswrapper[4868]: I1201 17:51:09.232279 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qdlmm" event={"ID":"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f","Type":"ContainerDied","Data":"1d371a066df984ec7e71cb33824e8997f0417a05a70e89e69f04d0a5d1eacb13"} Dec 01 17:51:09 crc kubenswrapper[4868]: I1201 17:51:09.232308 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qdlmm" event={"ID":"4fe4c201-9b3b-4cfe-a020-48a47bac2e9f","Type":"ContainerDied","Data":"1beba41baf7abdb5452ca936bc050043e6448320b8a9bd8f58bcc6fc3ac049e3"} Dec 01 17:51:09 crc kubenswrapper[4868]: I1201 17:51:09.232326 4868 scope.go:117] "RemoveContainer" containerID="1d371a066df984ec7e71cb33824e8997f0417a05a70e89e69f04d0a5d1eacb13" Dec 01 17:51:09 crc kubenswrapper[4868]: I1201 17:51:09.232371 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qdlmm" Dec 01 17:51:09 crc kubenswrapper[4868]: I1201 17:51:09.259299 4868 scope.go:117] "RemoveContainer" containerID="641f76135e1595e6e842926fa8df13749a638d0ab7b38351907d56b3ad3c3596" Dec 01 17:51:09 crc kubenswrapper[4868]: I1201 17:51:09.279762 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qdlmm"] Dec 01 17:51:09 crc kubenswrapper[4868]: I1201 17:51:09.288222 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qdlmm"] Dec 01 17:51:09 crc kubenswrapper[4868]: I1201 17:51:09.303927 4868 scope.go:117] "RemoveContainer" containerID="b00eb3ceaa7834244a3a0d2d81cb06509d7eee8d5a8670af12932cb73d9fc00f" Dec 01 17:51:09 crc kubenswrapper[4868]: I1201 17:51:09.337668 4868 scope.go:117] "RemoveContainer" containerID="1d371a066df984ec7e71cb33824e8997f0417a05a70e89e69f04d0a5d1eacb13" Dec 01 17:51:09 crc kubenswrapper[4868]: E1201 17:51:09.338367 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d371a066df984ec7e71cb33824e8997f0417a05a70e89e69f04d0a5d1eacb13\": container with ID starting with 1d371a066df984ec7e71cb33824e8997f0417a05a70e89e69f04d0a5d1eacb13 not found: ID does not exist" containerID="1d371a066df984ec7e71cb33824e8997f0417a05a70e89e69f04d0a5d1eacb13" Dec 01 17:51:09 crc kubenswrapper[4868]: I1201 17:51:09.338494 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d371a066df984ec7e71cb33824e8997f0417a05a70e89e69f04d0a5d1eacb13"} err="failed to get container status \"1d371a066df984ec7e71cb33824e8997f0417a05a70e89e69f04d0a5d1eacb13\": rpc error: code = NotFound desc = could not find container \"1d371a066df984ec7e71cb33824e8997f0417a05a70e89e69f04d0a5d1eacb13\": container with ID starting with 1d371a066df984ec7e71cb33824e8997f0417a05a70e89e69f04d0a5d1eacb13 not found: ID does not exist" Dec 01 17:51:09 crc kubenswrapper[4868]: I1201 17:51:09.338598 4868 scope.go:117] "RemoveContainer" containerID="641f76135e1595e6e842926fa8df13749a638d0ab7b38351907d56b3ad3c3596" Dec 01 17:51:09 crc kubenswrapper[4868]: E1201 17:51:09.339128 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"641f76135e1595e6e842926fa8df13749a638d0ab7b38351907d56b3ad3c3596\": container with ID starting with 641f76135e1595e6e842926fa8df13749a638d0ab7b38351907d56b3ad3c3596 not found: ID does not exist" containerID="641f76135e1595e6e842926fa8df13749a638d0ab7b38351907d56b3ad3c3596" Dec 01 17:51:09 crc kubenswrapper[4868]: I1201 17:51:09.339230 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"641f76135e1595e6e842926fa8df13749a638d0ab7b38351907d56b3ad3c3596"} err="failed to get container status \"641f76135e1595e6e842926fa8df13749a638d0ab7b38351907d56b3ad3c3596\": rpc error: code = NotFound desc = could not find container \"641f76135e1595e6e842926fa8df13749a638d0ab7b38351907d56b3ad3c3596\": container with ID starting with 641f76135e1595e6e842926fa8df13749a638d0ab7b38351907d56b3ad3c3596 not found: ID does not exist" Dec 01 17:51:09 crc kubenswrapper[4868]: I1201 17:51:09.339311 4868 scope.go:117] "RemoveContainer" containerID="b00eb3ceaa7834244a3a0d2d81cb06509d7eee8d5a8670af12932cb73d9fc00f" Dec 01 17:51:09 crc kubenswrapper[4868]: E1201 17:51:09.339715 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b00eb3ceaa7834244a3a0d2d81cb06509d7eee8d5a8670af12932cb73d9fc00f\": container with ID starting with b00eb3ceaa7834244a3a0d2d81cb06509d7eee8d5a8670af12932cb73d9fc00f not found: ID does not exist" containerID="b00eb3ceaa7834244a3a0d2d81cb06509d7eee8d5a8670af12932cb73d9fc00f" Dec 01 17:51:09 crc kubenswrapper[4868]: I1201 17:51:09.339761 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b00eb3ceaa7834244a3a0d2d81cb06509d7eee8d5a8670af12932cb73d9fc00f"} err="failed to get container status \"b00eb3ceaa7834244a3a0d2d81cb06509d7eee8d5a8670af12932cb73d9fc00f\": rpc error: code = NotFound desc = could not find container \"b00eb3ceaa7834244a3a0d2d81cb06509d7eee8d5a8670af12932cb73d9fc00f\": container with ID starting with b00eb3ceaa7834244a3a0d2d81cb06509d7eee8d5a8670af12932cb73d9fc00f not found: ID does not exist" Dec 01 17:51:10 crc kubenswrapper[4868]: I1201 17:51:10.330693 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fe4c201-9b3b-4cfe-a020-48a47bac2e9f" path="/var/lib/kubelet/pods/4fe4c201-9b3b-4cfe-a020-48a47bac2e9f/volumes" Dec 01 17:51:14 crc kubenswrapper[4868]: I1201 17:51:14.331507 4868 generic.go:334] "Generic (PLEG): container finished" podID="af68e45f-ae90-4a6b-a189-7a97b42d0b95" containerID="b0f54559e1315b65f72c499bf86eaf341d919617b878f84b87771d6e3ed337b1" exitCode=0 Dec 01 17:51:14 crc kubenswrapper[4868]: I1201 17:51:14.331593 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" event={"ID":"af68e45f-ae90-4a6b-a189-7a97b42d0b95","Type":"ContainerDied","Data":"b0f54559e1315b65f72c499bf86eaf341d919617b878f84b87771d6e3ed337b1"} Dec 01 17:51:15 crc kubenswrapper[4868]: I1201 17:51:15.822273 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:51:15 crc kubenswrapper[4868]: I1201 17:51:15.966639 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdq2j\" (UniqueName: \"kubernetes.io/projected/af68e45f-ae90-4a6b-a189-7a97b42d0b95-kube-api-access-xdq2j\") pod \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " Dec 01 17:51:15 crc kubenswrapper[4868]: I1201 17:51:15.966736 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-ssh-key\") pod \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " Dec 01 17:51:15 crc kubenswrapper[4868]: I1201 17:51:15.966836 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-inventory\") pod \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " Dec 01 17:51:15 crc kubenswrapper[4868]: I1201 17:51:15.966989 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-bootstrap-combined-ca-bundle\") pod \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\" (UID: \"af68e45f-ae90-4a6b-a189-7a97b42d0b95\") " Dec 01 17:51:15 crc kubenswrapper[4868]: I1201 17:51:15.979181 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "af68e45f-ae90-4a6b-a189-7a97b42d0b95" (UID: "af68e45f-ae90-4a6b-a189-7a97b42d0b95"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:51:15 crc kubenswrapper[4868]: I1201 17:51:15.979186 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af68e45f-ae90-4a6b-a189-7a97b42d0b95-kube-api-access-xdq2j" (OuterVolumeSpecName: "kube-api-access-xdq2j") pod "af68e45f-ae90-4a6b-a189-7a97b42d0b95" (UID: "af68e45f-ae90-4a6b-a189-7a97b42d0b95"). InnerVolumeSpecName "kube-api-access-xdq2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.003809 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-inventory" (OuterVolumeSpecName: "inventory") pod "af68e45f-ae90-4a6b-a189-7a97b42d0b95" (UID: "af68e45f-ae90-4a6b-a189-7a97b42d0b95"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.006118 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "af68e45f-ae90-4a6b-a189-7a97b42d0b95" (UID: "af68e45f-ae90-4a6b-a189-7a97b42d0b95"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.068862 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdq2j\" (UniqueName: \"kubernetes.io/projected/af68e45f-ae90-4a6b-a189-7a97b42d0b95-kube-api-access-xdq2j\") on node \"crc\" DevicePath \"\"" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.068899 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.068911 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.068926 4868 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af68e45f-ae90-4a6b-a189-7a97b42d0b95-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.357759 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" event={"ID":"af68e45f-ae90-4a6b-a189-7a97b42d0b95","Type":"ContainerDied","Data":"a916122ca7b4daa94a39a28317643c51842241ceabd9b123b9c5b6c9ebc13430"} Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.357807 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a916122ca7b4daa94a39a28317643c51842241ceabd9b123b9c5b6c9ebc13430" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.357846 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.441431 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b"] Dec 01 17:51:16 crc kubenswrapper[4868]: E1201 17:51:16.441880 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af68e45f-ae90-4a6b-a189-7a97b42d0b95" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.441900 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="af68e45f-ae90-4a6b-a189-7a97b42d0b95" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 17:51:16 crc kubenswrapper[4868]: E1201 17:51:16.441916 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fe4c201-9b3b-4cfe-a020-48a47bac2e9f" containerName="extract-content" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.441923 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fe4c201-9b3b-4cfe-a020-48a47bac2e9f" containerName="extract-content" Dec 01 17:51:16 crc kubenswrapper[4868]: E1201 17:51:16.441932 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fe4c201-9b3b-4cfe-a020-48a47bac2e9f" containerName="registry-server" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.441992 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fe4c201-9b3b-4cfe-a020-48a47bac2e9f" containerName="registry-server" Dec 01 17:51:16 crc kubenswrapper[4868]: E1201 17:51:16.442037 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fe4c201-9b3b-4cfe-a020-48a47bac2e9f" containerName="extract-utilities" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.442045 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fe4c201-9b3b-4cfe-a020-48a47bac2e9f" containerName="extract-utilities" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.442262 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="af68e45f-ae90-4a6b-a189-7a97b42d0b95" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.442280 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fe4c201-9b3b-4cfe-a020-48a47bac2e9f" containerName="registry-server" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.442915 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.445372 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.445526 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.445646 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.446843 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.452834 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b"] Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.579745 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e39e56d-e163-4e75-909f-410c72fde805-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b\" (UID: \"6e39e56d-e163-4e75-909f-410c72fde805\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.579835 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e39e56d-e163-4e75-909f-410c72fde805-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b\" (UID: \"6e39e56d-e163-4e75-909f-410c72fde805\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.580007 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b68pc\" (UniqueName: \"kubernetes.io/projected/6e39e56d-e163-4e75-909f-410c72fde805-kube-api-access-b68pc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b\" (UID: \"6e39e56d-e163-4e75-909f-410c72fde805\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.681232 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e39e56d-e163-4e75-909f-410c72fde805-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b\" (UID: \"6e39e56d-e163-4e75-909f-410c72fde805\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.681358 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b68pc\" (UniqueName: \"kubernetes.io/projected/6e39e56d-e163-4e75-909f-410c72fde805-kube-api-access-b68pc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b\" (UID: \"6e39e56d-e163-4e75-909f-410c72fde805\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.681411 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e39e56d-e163-4e75-909f-410c72fde805-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b\" (UID: \"6e39e56d-e163-4e75-909f-410c72fde805\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.689113 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e39e56d-e163-4e75-909f-410c72fde805-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b\" (UID: \"6e39e56d-e163-4e75-909f-410c72fde805\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.693873 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e39e56d-e163-4e75-909f-410c72fde805-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b\" (UID: \"6e39e56d-e163-4e75-909f-410c72fde805\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.698890 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b68pc\" (UniqueName: \"kubernetes.io/projected/6e39e56d-e163-4e75-909f-410c72fde805-kube-api-access-b68pc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b\" (UID: \"6e39e56d-e163-4e75-909f-410c72fde805\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" Dec 01 17:51:16 crc kubenswrapper[4868]: I1201 17:51:16.765014 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" Dec 01 17:51:17 crc kubenswrapper[4868]: I1201 17:51:17.295536 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b"] Dec 01 17:51:17 crc kubenswrapper[4868]: I1201 17:51:17.306671 4868 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 17:51:17 crc kubenswrapper[4868]: I1201 17:51:17.367388 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" event={"ID":"6e39e56d-e163-4e75-909f-410c72fde805","Type":"ContainerStarted","Data":"462de88fbde64e2a3b81bdc38090df6fa08aa2448dfa85611db5bc1b91969538"} Dec 01 17:51:18 crc kubenswrapper[4868]: I1201 17:51:18.377990 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" event={"ID":"6e39e56d-e163-4e75-909f-410c72fde805","Type":"ContainerStarted","Data":"ef99afb01e648b4610541c68c101aa006cb8dd1b51cf387a4704a5e23a46f9c8"} Dec 01 17:51:18 crc kubenswrapper[4868]: I1201 17:51:18.404551 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" podStartSLOduration=1.672393072 podStartE2EDuration="2.404532706s" podCreationTimestamp="2025-12-01 17:51:16 +0000 UTC" firstStartedPulling="2025-12-01 17:51:17.306448733 +0000 UTC m=+1549.677559144" lastFinishedPulling="2025-12-01 17:51:18.038588367 +0000 UTC m=+1550.409698778" observedRunningTime="2025-12-01 17:51:18.396362884 +0000 UTC m=+1550.767473295" watchObservedRunningTime="2025-12-01 17:51:18.404532706 +0000 UTC m=+1550.775643117" Dec 01 17:51:23 crc kubenswrapper[4868]: I1201 17:51:23.176826 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-znz8b"] Dec 01 17:51:23 crc kubenswrapper[4868]: I1201 17:51:23.180543 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:23 crc kubenswrapper[4868]: I1201 17:51:23.186441 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-znz8b"] Dec 01 17:51:23 crc kubenswrapper[4868]: I1201 17:51:23.303965 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk4s8\" (UniqueName: \"kubernetes.io/projected/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-kube-api-access-rk4s8\") pod \"redhat-marketplace-znz8b\" (UID: \"1278bdd0-d037-4fc4-9d1c-02e0b6a83952\") " pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:23 crc kubenswrapper[4868]: I1201 17:51:23.304028 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-catalog-content\") pod \"redhat-marketplace-znz8b\" (UID: \"1278bdd0-d037-4fc4-9d1c-02e0b6a83952\") " pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:23 crc kubenswrapper[4868]: I1201 17:51:23.304520 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-utilities\") pod \"redhat-marketplace-znz8b\" (UID: \"1278bdd0-d037-4fc4-9d1c-02e0b6a83952\") " pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:23 crc kubenswrapper[4868]: I1201 17:51:23.408795 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-utilities\") pod \"redhat-marketplace-znz8b\" (UID: \"1278bdd0-d037-4fc4-9d1c-02e0b6a83952\") " pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:23 crc kubenswrapper[4868]: I1201 17:51:23.408931 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk4s8\" (UniqueName: \"kubernetes.io/projected/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-kube-api-access-rk4s8\") pod \"redhat-marketplace-znz8b\" (UID: \"1278bdd0-d037-4fc4-9d1c-02e0b6a83952\") " pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:23 crc kubenswrapper[4868]: I1201 17:51:23.409101 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-catalog-content\") pod \"redhat-marketplace-znz8b\" (UID: \"1278bdd0-d037-4fc4-9d1c-02e0b6a83952\") " pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:23 crc kubenswrapper[4868]: I1201 17:51:23.409696 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-utilities\") pod \"redhat-marketplace-znz8b\" (UID: \"1278bdd0-d037-4fc4-9d1c-02e0b6a83952\") " pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:23 crc kubenswrapper[4868]: I1201 17:51:23.409939 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-catalog-content\") pod \"redhat-marketplace-znz8b\" (UID: \"1278bdd0-d037-4fc4-9d1c-02e0b6a83952\") " pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:23 crc kubenswrapper[4868]: I1201 17:51:23.432438 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk4s8\" (UniqueName: \"kubernetes.io/projected/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-kube-api-access-rk4s8\") pod \"redhat-marketplace-znz8b\" (UID: \"1278bdd0-d037-4fc4-9d1c-02e0b6a83952\") " pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:23 crc kubenswrapper[4868]: I1201 17:51:23.503468 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:24 crc kubenswrapper[4868]: I1201 17:51:24.050040 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-znz8b"] Dec 01 17:51:24 crc kubenswrapper[4868]: I1201 17:51:24.455934 4868 generic.go:334] "Generic (PLEG): container finished" podID="1278bdd0-d037-4fc4-9d1c-02e0b6a83952" containerID="da395b7b6a2ad62722ce215797f93af11d5d430c560d97dc5af2ab29dbef2c3b" exitCode=0 Dec 01 17:51:24 crc kubenswrapper[4868]: I1201 17:51:24.456468 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-znz8b" event={"ID":"1278bdd0-d037-4fc4-9d1c-02e0b6a83952","Type":"ContainerDied","Data":"da395b7b6a2ad62722ce215797f93af11d5d430c560d97dc5af2ab29dbef2c3b"} Dec 01 17:51:24 crc kubenswrapper[4868]: I1201 17:51:24.456544 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-znz8b" event={"ID":"1278bdd0-d037-4fc4-9d1c-02e0b6a83952","Type":"ContainerStarted","Data":"44df7a274df1b31d6b6539e91fad7aab044bf913cb7d8489e8eb449f798e6d63"} Dec 01 17:51:26 crc kubenswrapper[4868]: I1201 17:51:26.484312 4868 generic.go:334] "Generic (PLEG): container finished" podID="1278bdd0-d037-4fc4-9d1c-02e0b6a83952" containerID="8263111d41eef9887294197b139e021634a9d63d7d05012883a72cbe3ca003db" exitCode=0 Dec 01 17:51:26 crc kubenswrapper[4868]: I1201 17:51:26.484403 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-znz8b" event={"ID":"1278bdd0-d037-4fc4-9d1c-02e0b6a83952","Type":"ContainerDied","Data":"8263111d41eef9887294197b139e021634a9d63d7d05012883a72cbe3ca003db"} Dec 01 17:51:27 crc kubenswrapper[4868]: I1201 17:51:27.497435 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-znz8b" event={"ID":"1278bdd0-d037-4fc4-9d1c-02e0b6a83952","Type":"ContainerStarted","Data":"929d1293748d913d0bdab3c335d0cc66490330c324628aaf26ff684495095826"} Dec 01 17:51:33 crc kubenswrapper[4868]: I1201 17:51:33.503984 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:33 crc kubenswrapper[4868]: I1201 17:51:33.504524 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:33 crc kubenswrapper[4868]: I1201 17:51:33.546842 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:33 crc kubenswrapper[4868]: I1201 17:51:33.570123 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-znz8b" podStartSLOduration=8.010913808 podStartE2EDuration="10.570104584s" podCreationTimestamp="2025-12-01 17:51:23 +0000 UTC" firstStartedPulling="2025-12-01 17:51:24.457860802 +0000 UTC m=+1556.828971213" lastFinishedPulling="2025-12-01 17:51:27.017051578 +0000 UTC m=+1559.388161989" observedRunningTime="2025-12-01 17:51:27.516187601 +0000 UTC m=+1559.887298032" watchObservedRunningTime="2025-12-01 17:51:33.570104584 +0000 UTC m=+1565.941214995" Dec 01 17:51:33 crc kubenswrapper[4868]: I1201 17:51:33.606078 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:33 crc kubenswrapper[4868]: I1201 17:51:33.788185 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-znz8b"] Dec 01 17:51:35 crc kubenswrapper[4868]: I1201 17:51:35.568202 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-znz8b" podUID="1278bdd0-d037-4fc4-9d1c-02e0b6a83952" containerName="registry-server" containerID="cri-o://929d1293748d913d0bdab3c335d0cc66490330c324628aaf26ff684495095826" gracePeriod=2 Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.062059 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.181065 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-catalog-content\") pod \"1278bdd0-d037-4fc4-9d1c-02e0b6a83952\" (UID: \"1278bdd0-d037-4fc4-9d1c-02e0b6a83952\") " Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.181271 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rk4s8\" (UniqueName: \"kubernetes.io/projected/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-kube-api-access-rk4s8\") pod \"1278bdd0-d037-4fc4-9d1c-02e0b6a83952\" (UID: \"1278bdd0-d037-4fc4-9d1c-02e0b6a83952\") " Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.181382 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-utilities\") pod \"1278bdd0-d037-4fc4-9d1c-02e0b6a83952\" (UID: \"1278bdd0-d037-4fc4-9d1c-02e0b6a83952\") " Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.182623 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-utilities" (OuterVolumeSpecName: "utilities") pod "1278bdd0-d037-4fc4-9d1c-02e0b6a83952" (UID: "1278bdd0-d037-4fc4-9d1c-02e0b6a83952"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.187073 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-kube-api-access-rk4s8" (OuterVolumeSpecName: "kube-api-access-rk4s8") pod "1278bdd0-d037-4fc4-9d1c-02e0b6a83952" (UID: "1278bdd0-d037-4fc4-9d1c-02e0b6a83952"). InnerVolumeSpecName "kube-api-access-rk4s8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.200966 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1278bdd0-d037-4fc4-9d1c-02e0b6a83952" (UID: "1278bdd0-d037-4fc4-9d1c-02e0b6a83952"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.283628 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rk4s8\" (UniqueName: \"kubernetes.io/projected/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-kube-api-access-rk4s8\") on node \"crc\" DevicePath \"\"" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.284026 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.284045 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1278bdd0-d037-4fc4-9d1c-02e0b6a83952-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.587514 4868 generic.go:334] "Generic (PLEG): container finished" podID="1278bdd0-d037-4fc4-9d1c-02e0b6a83952" containerID="929d1293748d913d0bdab3c335d0cc66490330c324628aaf26ff684495095826" exitCode=0 Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.587633 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-znz8b" event={"ID":"1278bdd0-d037-4fc4-9d1c-02e0b6a83952","Type":"ContainerDied","Data":"929d1293748d913d0bdab3c335d0cc66490330c324628aaf26ff684495095826"} Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.587728 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-znz8b" event={"ID":"1278bdd0-d037-4fc4-9d1c-02e0b6a83952","Type":"ContainerDied","Data":"44df7a274df1b31d6b6539e91fad7aab044bf913cb7d8489e8eb449f798e6d63"} Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.587816 4868 scope.go:117] "RemoveContainer" containerID="929d1293748d913d0bdab3c335d0cc66490330c324628aaf26ff684495095826" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.588331 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-znz8b" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.633968 4868 scope.go:117] "RemoveContainer" containerID="8263111d41eef9887294197b139e021634a9d63d7d05012883a72cbe3ca003db" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.635735 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-znz8b"] Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.645393 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-znz8b"] Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.658353 4868 scope.go:117] "RemoveContainer" containerID="da395b7b6a2ad62722ce215797f93af11d5d430c560d97dc5af2ab29dbef2c3b" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.726076 4868 scope.go:117] "RemoveContainer" containerID="929d1293748d913d0bdab3c335d0cc66490330c324628aaf26ff684495095826" Dec 01 17:51:36 crc kubenswrapper[4868]: E1201 17:51:36.726597 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"929d1293748d913d0bdab3c335d0cc66490330c324628aaf26ff684495095826\": container with ID starting with 929d1293748d913d0bdab3c335d0cc66490330c324628aaf26ff684495095826 not found: ID does not exist" containerID="929d1293748d913d0bdab3c335d0cc66490330c324628aaf26ff684495095826" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.726641 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"929d1293748d913d0bdab3c335d0cc66490330c324628aaf26ff684495095826"} err="failed to get container status \"929d1293748d913d0bdab3c335d0cc66490330c324628aaf26ff684495095826\": rpc error: code = NotFound desc = could not find container \"929d1293748d913d0bdab3c335d0cc66490330c324628aaf26ff684495095826\": container with ID starting with 929d1293748d913d0bdab3c335d0cc66490330c324628aaf26ff684495095826 not found: ID does not exist" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.726668 4868 scope.go:117] "RemoveContainer" containerID="8263111d41eef9887294197b139e021634a9d63d7d05012883a72cbe3ca003db" Dec 01 17:51:36 crc kubenswrapper[4868]: E1201 17:51:36.727010 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8263111d41eef9887294197b139e021634a9d63d7d05012883a72cbe3ca003db\": container with ID starting with 8263111d41eef9887294197b139e021634a9d63d7d05012883a72cbe3ca003db not found: ID does not exist" containerID="8263111d41eef9887294197b139e021634a9d63d7d05012883a72cbe3ca003db" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.727034 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8263111d41eef9887294197b139e021634a9d63d7d05012883a72cbe3ca003db"} err="failed to get container status \"8263111d41eef9887294197b139e021634a9d63d7d05012883a72cbe3ca003db\": rpc error: code = NotFound desc = could not find container \"8263111d41eef9887294197b139e021634a9d63d7d05012883a72cbe3ca003db\": container with ID starting with 8263111d41eef9887294197b139e021634a9d63d7d05012883a72cbe3ca003db not found: ID does not exist" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.727053 4868 scope.go:117] "RemoveContainer" containerID="da395b7b6a2ad62722ce215797f93af11d5d430c560d97dc5af2ab29dbef2c3b" Dec 01 17:51:36 crc kubenswrapper[4868]: E1201 17:51:36.727275 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da395b7b6a2ad62722ce215797f93af11d5d430c560d97dc5af2ab29dbef2c3b\": container with ID starting with da395b7b6a2ad62722ce215797f93af11d5d430c560d97dc5af2ab29dbef2c3b not found: ID does not exist" containerID="da395b7b6a2ad62722ce215797f93af11d5d430c560d97dc5af2ab29dbef2c3b" Dec 01 17:51:36 crc kubenswrapper[4868]: I1201 17:51:36.727293 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da395b7b6a2ad62722ce215797f93af11d5d430c560d97dc5af2ab29dbef2c3b"} err="failed to get container status \"da395b7b6a2ad62722ce215797f93af11d5d430c560d97dc5af2ab29dbef2c3b\": rpc error: code = NotFound desc = could not find container \"da395b7b6a2ad62722ce215797f93af11d5d430c560d97dc5af2ab29dbef2c3b\": container with ID starting with da395b7b6a2ad62722ce215797f93af11d5d430c560d97dc5af2ab29dbef2c3b not found: ID does not exist" Dec 01 17:51:38 crc kubenswrapper[4868]: I1201 17:51:38.184165 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1278bdd0-d037-4fc4-9d1c-02e0b6a83952" path="/var/lib/kubelet/pods/1278bdd0-d037-4fc4-9d1c-02e0b6a83952/volumes" Dec 01 17:51:38 crc kubenswrapper[4868]: I1201 17:51:38.956863 4868 scope.go:117] "RemoveContainer" containerID="68c50850b354708778d1c5f0bf21257e977a75918b3a2edeb8bcb5d0ae9e3f23" Dec 01 17:51:38 crc kubenswrapper[4868]: I1201 17:51:38.992207 4868 scope.go:117] "RemoveContainer" containerID="f15b7bfacdd6842acbe87c872c21527f51a023616f5ae43cfb4af45011ae642e" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.482957 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jg5m2"] Dec 01 17:51:47 crc kubenswrapper[4868]: E1201 17:51:47.483994 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1278bdd0-d037-4fc4-9d1c-02e0b6a83952" containerName="extract-utilities" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.484010 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="1278bdd0-d037-4fc4-9d1c-02e0b6a83952" containerName="extract-utilities" Dec 01 17:51:47 crc kubenswrapper[4868]: E1201 17:51:47.484020 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1278bdd0-d037-4fc4-9d1c-02e0b6a83952" containerName="registry-server" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.484027 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="1278bdd0-d037-4fc4-9d1c-02e0b6a83952" containerName="registry-server" Dec 01 17:51:47 crc kubenswrapper[4868]: E1201 17:51:47.484052 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1278bdd0-d037-4fc4-9d1c-02e0b6a83952" containerName="extract-content" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.484058 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="1278bdd0-d037-4fc4-9d1c-02e0b6a83952" containerName="extract-content" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.484244 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="1278bdd0-d037-4fc4-9d1c-02e0b6a83952" containerName="registry-server" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.485743 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.502313 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jg5m2"] Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.628458 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37079409-9062-477e-98a5-ca24163cef35-utilities\") pod \"redhat-operators-jg5m2\" (UID: \"37079409-9062-477e-98a5-ca24163cef35\") " pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.628629 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l78bl\" (UniqueName: \"kubernetes.io/projected/37079409-9062-477e-98a5-ca24163cef35-kube-api-access-l78bl\") pod \"redhat-operators-jg5m2\" (UID: \"37079409-9062-477e-98a5-ca24163cef35\") " pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.628723 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37079409-9062-477e-98a5-ca24163cef35-catalog-content\") pod \"redhat-operators-jg5m2\" (UID: \"37079409-9062-477e-98a5-ca24163cef35\") " pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.730722 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37079409-9062-477e-98a5-ca24163cef35-catalog-content\") pod \"redhat-operators-jg5m2\" (UID: \"37079409-9062-477e-98a5-ca24163cef35\") " pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.730837 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37079409-9062-477e-98a5-ca24163cef35-utilities\") pod \"redhat-operators-jg5m2\" (UID: \"37079409-9062-477e-98a5-ca24163cef35\") " pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.730992 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l78bl\" (UniqueName: \"kubernetes.io/projected/37079409-9062-477e-98a5-ca24163cef35-kube-api-access-l78bl\") pod \"redhat-operators-jg5m2\" (UID: \"37079409-9062-477e-98a5-ca24163cef35\") " pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.731380 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37079409-9062-477e-98a5-ca24163cef35-catalog-content\") pod \"redhat-operators-jg5m2\" (UID: \"37079409-9062-477e-98a5-ca24163cef35\") " pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.731443 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37079409-9062-477e-98a5-ca24163cef35-utilities\") pod \"redhat-operators-jg5m2\" (UID: \"37079409-9062-477e-98a5-ca24163cef35\") " pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.753054 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l78bl\" (UniqueName: \"kubernetes.io/projected/37079409-9062-477e-98a5-ca24163cef35-kube-api-access-l78bl\") pod \"redhat-operators-jg5m2\" (UID: \"37079409-9062-477e-98a5-ca24163cef35\") " pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:51:47 crc kubenswrapper[4868]: I1201 17:51:47.808815 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:51:48 crc kubenswrapper[4868]: I1201 17:51:48.285788 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jg5m2"] Dec 01 17:51:48 crc kubenswrapper[4868]: I1201 17:51:48.708641 4868 generic.go:334] "Generic (PLEG): container finished" podID="37079409-9062-477e-98a5-ca24163cef35" containerID="74dbf98cb4039c6b9a186b99ad8458dee86bf901adb84ce5f6913bc2eee3e3ed" exitCode=0 Dec 01 17:51:48 crc kubenswrapper[4868]: I1201 17:51:48.708739 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jg5m2" event={"ID":"37079409-9062-477e-98a5-ca24163cef35","Type":"ContainerDied","Data":"74dbf98cb4039c6b9a186b99ad8458dee86bf901adb84ce5f6913bc2eee3e3ed"} Dec 01 17:51:48 crc kubenswrapper[4868]: I1201 17:51:48.709221 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jg5m2" event={"ID":"37079409-9062-477e-98a5-ca24163cef35","Type":"ContainerStarted","Data":"33a80248a66eb59a27fbe0f1f675b06fee4ea23f827e29a87285191dca57913d"} Dec 01 17:51:50 crc kubenswrapper[4868]: I1201 17:51:50.731219 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jg5m2" event={"ID":"37079409-9062-477e-98a5-ca24163cef35","Type":"ContainerStarted","Data":"0d1270fdf53d5be1b36153140f916a6f521ae0acc13d71dc522c4b0008c74223"} Dec 01 17:51:53 crc kubenswrapper[4868]: I1201 17:51:53.757299 4868 generic.go:334] "Generic (PLEG): container finished" podID="37079409-9062-477e-98a5-ca24163cef35" containerID="0d1270fdf53d5be1b36153140f916a6f521ae0acc13d71dc522c4b0008c74223" exitCode=0 Dec 01 17:51:53 crc kubenswrapper[4868]: I1201 17:51:53.757767 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jg5m2" event={"ID":"37079409-9062-477e-98a5-ca24163cef35","Type":"ContainerDied","Data":"0d1270fdf53d5be1b36153140f916a6f521ae0acc13d71dc522c4b0008c74223"} Dec 01 17:51:54 crc kubenswrapper[4868]: I1201 17:51:54.771788 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jg5m2" event={"ID":"37079409-9062-477e-98a5-ca24163cef35","Type":"ContainerStarted","Data":"75a67fb72b0e9b2b69507740e4b8a433a0f6fe1b075dd732bb8cfbe6b7909a1d"} Dec 01 17:51:54 crc kubenswrapper[4868]: I1201 17:51:54.801869 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jg5m2" podStartSLOduration=2.285891247 podStartE2EDuration="7.801829148s" podCreationTimestamp="2025-12-01 17:51:47 +0000 UTC" firstStartedPulling="2025-12-01 17:51:48.710594771 +0000 UTC m=+1581.081705182" lastFinishedPulling="2025-12-01 17:51:54.226532682 +0000 UTC m=+1586.597643083" observedRunningTime="2025-12-01 17:51:54.793631665 +0000 UTC m=+1587.164742086" watchObservedRunningTime="2025-12-01 17:51:54.801829148 +0000 UTC m=+1587.172939559" Dec 01 17:51:57 crc kubenswrapper[4868]: I1201 17:51:57.809452 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:51:57 crc kubenswrapper[4868]: I1201 17:51:57.809745 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:51:58 crc kubenswrapper[4868]: I1201 17:51:58.857840 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jg5m2" podUID="37079409-9062-477e-98a5-ca24163cef35" containerName="registry-server" probeResult="failure" output=< Dec 01 17:51:58 crc kubenswrapper[4868]: timeout: failed to connect service ":50051" within 1s Dec 01 17:51:58 crc kubenswrapper[4868]: > Dec 01 17:52:07 crc kubenswrapper[4868]: I1201 17:52:07.867526 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:52:07 crc kubenswrapper[4868]: I1201 17:52:07.914229 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:52:08 crc kubenswrapper[4868]: I1201 17:52:08.100115 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jg5m2"] Dec 01 17:52:08 crc kubenswrapper[4868]: I1201 17:52:08.903553 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jg5m2" podUID="37079409-9062-477e-98a5-ca24163cef35" containerName="registry-server" containerID="cri-o://75a67fb72b0e9b2b69507740e4b8a433a0f6fe1b075dd732bb8cfbe6b7909a1d" gracePeriod=2 Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.349792 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.479770 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37079409-9062-477e-98a5-ca24163cef35-catalog-content\") pod \"37079409-9062-477e-98a5-ca24163cef35\" (UID: \"37079409-9062-477e-98a5-ca24163cef35\") " Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.479870 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37079409-9062-477e-98a5-ca24163cef35-utilities\") pod \"37079409-9062-477e-98a5-ca24163cef35\" (UID: \"37079409-9062-477e-98a5-ca24163cef35\") " Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.480081 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l78bl\" (UniqueName: \"kubernetes.io/projected/37079409-9062-477e-98a5-ca24163cef35-kube-api-access-l78bl\") pod \"37079409-9062-477e-98a5-ca24163cef35\" (UID: \"37079409-9062-477e-98a5-ca24163cef35\") " Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.481022 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37079409-9062-477e-98a5-ca24163cef35-utilities" (OuterVolumeSpecName: "utilities") pod "37079409-9062-477e-98a5-ca24163cef35" (UID: "37079409-9062-477e-98a5-ca24163cef35"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.491385 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37079409-9062-477e-98a5-ca24163cef35-kube-api-access-l78bl" (OuterVolumeSpecName: "kube-api-access-l78bl") pod "37079409-9062-477e-98a5-ca24163cef35" (UID: "37079409-9062-477e-98a5-ca24163cef35"). InnerVolumeSpecName "kube-api-access-l78bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.582071 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l78bl\" (UniqueName: \"kubernetes.io/projected/37079409-9062-477e-98a5-ca24163cef35-kube-api-access-l78bl\") on node \"crc\" DevicePath \"\"" Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.582105 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37079409-9062-477e-98a5-ca24163cef35-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.596270 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37079409-9062-477e-98a5-ca24163cef35-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37079409-9062-477e-98a5-ca24163cef35" (UID: "37079409-9062-477e-98a5-ca24163cef35"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.684106 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37079409-9062-477e-98a5-ca24163cef35-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.913657 4868 generic.go:334] "Generic (PLEG): container finished" podID="37079409-9062-477e-98a5-ca24163cef35" containerID="75a67fb72b0e9b2b69507740e4b8a433a0f6fe1b075dd732bb8cfbe6b7909a1d" exitCode=0 Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.913717 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jg5m2" Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.913732 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jg5m2" event={"ID":"37079409-9062-477e-98a5-ca24163cef35","Type":"ContainerDied","Data":"75a67fb72b0e9b2b69507740e4b8a433a0f6fe1b075dd732bb8cfbe6b7909a1d"} Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.913794 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jg5m2" event={"ID":"37079409-9062-477e-98a5-ca24163cef35","Type":"ContainerDied","Data":"33a80248a66eb59a27fbe0f1f675b06fee4ea23f827e29a87285191dca57913d"} Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.913817 4868 scope.go:117] "RemoveContainer" containerID="75a67fb72b0e9b2b69507740e4b8a433a0f6fe1b075dd732bb8cfbe6b7909a1d" Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.946591 4868 scope.go:117] "RemoveContainer" containerID="0d1270fdf53d5be1b36153140f916a6f521ae0acc13d71dc522c4b0008c74223" Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.950743 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jg5m2"] Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.960256 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jg5m2"] Dec 01 17:52:09 crc kubenswrapper[4868]: I1201 17:52:09.974035 4868 scope.go:117] "RemoveContainer" containerID="74dbf98cb4039c6b9a186b99ad8458dee86bf901adb84ce5f6913bc2eee3e3ed" Dec 01 17:52:10 crc kubenswrapper[4868]: I1201 17:52:10.013657 4868 scope.go:117] "RemoveContainer" containerID="75a67fb72b0e9b2b69507740e4b8a433a0f6fe1b075dd732bb8cfbe6b7909a1d" Dec 01 17:52:10 crc kubenswrapper[4868]: E1201 17:52:10.014471 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75a67fb72b0e9b2b69507740e4b8a433a0f6fe1b075dd732bb8cfbe6b7909a1d\": container with ID starting with 75a67fb72b0e9b2b69507740e4b8a433a0f6fe1b075dd732bb8cfbe6b7909a1d not found: ID does not exist" containerID="75a67fb72b0e9b2b69507740e4b8a433a0f6fe1b075dd732bb8cfbe6b7909a1d" Dec 01 17:52:10 crc kubenswrapper[4868]: I1201 17:52:10.014560 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75a67fb72b0e9b2b69507740e4b8a433a0f6fe1b075dd732bb8cfbe6b7909a1d"} err="failed to get container status \"75a67fb72b0e9b2b69507740e4b8a433a0f6fe1b075dd732bb8cfbe6b7909a1d\": rpc error: code = NotFound desc = could not find container \"75a67fb72b0e9b2b69507740e4b8a433a0f6fe1b075dd732bb8cfbe6b7909a1d\": container with ID starting with 75a67fb72b0e9b2b69507740e4b8a433a0f6fe1b075dd732bb8cfbe6b7909a1d not found: ID does not exist" Dec 01 17:52:10 crc kubenswrapper[4868]: I1201 17:52:10.014702 4868 scope.go:117] "RemoveContainer" containerID="0d1270fdf53d5be1b36153140f916a6f521ae0acc13d71dc522c4b0008c74223" Dec 01 17:52:10 crc kubenswrapper[4868]: E1201 17:52:10.015205 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d1270fdf53d5be1b36153140f916a6f521ae0acc13d71dc522c4b0008c74223\": container with ID starting with 0d1270fdf53d5be1b36153140f916a6f521ae0acc13d71dc522c4b0008c74223 not found: ID does not exist" containerID="0d1270fdf53d5be1b36153140f916a6f521ae0acc13d71dc522c4b0008c74223" Dec 01 17:52:10 crc kubenswrapper[4868]: I1201 17:52:10.015238 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d1270fdf53d5be1b36153140f916a6f521ae0acc13d71dc522c4b0008c74223"} err="failed to get container status \"0d1270fdf53d5be1b36153140f916a6f521ae0acc13d71dc522c4b0008c74223\": rpc error: code = NotFound desc = could not find container \"0d1270fdf53d5be1b36153140f916a6f521ae0acc13d71dc522c4b0008c74223\": container with ID starting with 0d1270fdf53d5be1b36153140f916a6f521ae0acc13d71dc522c4b0008c74223 not found: ID does not exist" Dec 01 17:52:10 crc kubenswrapper[4868]: I1201 17:52:10.015262 4868 scope.go:117] "RemoveContainer" containerID="74dbf98cb4039c6b9a186b99ad8458dee86bf901adb84ce5f6913bc2eee3e3ed" Dec 01 17:52:10 crc kubenswrapper[4868]: E1201 17:52:10.015597 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74dbf98cb4039c6b9a186b99ad8458dee86bf901adb84ce5f6913bc2eee3e3ed\": container with ID starting with 74dbf98cb4039c6b9a186b99ad8458dee86bf901adb84ce5f6913bc2eee3e3ed not found: ID does not exist" containerID="74dbf98cb4039c6b9a186b99ad8458dee86bf901adb84ce5f6913bc2eee3e3ed" Dec 01 17:52:10 crc kubenswrapper[4868]: I1201 17:52:10.015643 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74dbf98cb4039c6b9a186b99ad8458dee86bf901adb84ce5f6913bc2eee3e3ed"} err="failed to get container status \"74dbf98cb4039c6b9a186b99ad8458dee86bf901adb84ce5f6913bc2eee3e3ed\": rpc error: code = NotFound desc = could not find container \"74dbf98cb4039c6b9a186b99ad8458dee86bf901adb84ce5f6913bc2eee3e3ed\": container with ID starting with 74dbf98cb4039c6b9a186b99ad8458dee86bf901adb84ce5f6913bc2eee3e3ed not found: ID does not exist" Dec 01 17:52:10 crc kubenswrapper[4868]: I1201 17:52:10.182280 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37079409-9062-477e-98a5-ca24163cef35" path="/var/lib/kubelet/pods/37079409-9062-477e-98a5-ca24163cef35/volumes" Dec 01 17:52:14 crc kubenswrapper[4868]: I1201 17:52:14.036629 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-592xb"] Dec 01 17:52:14 crc kubenswrapper[4868]: I1201 17:52:14.046535 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-c223-account-create-update-g7wbx"] Dec 01 17:52:14 crc kubenswrapper[4868]: I1201 17:52:14.056149 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-c223-account-create-update-g7wbx"] Dec 01 17:52:14 crc kubenswrapper[4868]: I1201 17:52:14.064686 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-592xb"] Dec 01 17:52:14 crc kubenswrapper[4868]: I1201 17:52:14.185357 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f1127ca-d13f-46cf-b792-f9854e37586c" path="/var/lib/kubelet/pods/0f1127ca-d13f-46cf-b792-f9854e37586c/volumes" Dec 01 17:52:14 crc kubenswrapper[4868]: I1201 17:52:14.186379 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f698ec2-ef7a-4614-8087-5b4f3fe94365" path="/var/lib/kubelet/pods/7f698ec2-ef7a-4614-8087-5b4f3fe94365/volumes" Dec 01 17:52:17 crc kubenswrapper[4868]: I1201 17:52:17.035315 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-6cp5w"] Dec 01 17:52:17 crc kubenswrapper[4868]: I1201 17:52:17.050464 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-1f4f-account-create-update-4qr5v"] Dec 01 17:52:17 crc kubenswrapper[4868]: I1201 17:52:17.062506 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-6cp5w"] Dec 01 17:52:17 crc kubenswrapper[4868]: I1201 17:52:17.074608 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-1f4f-account-create-update-4qr5v"] Dec 01 17:52:18 crc kubenswrapper[4868]: I1201 17:52:18.031074 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-8qflm"] Dec 01 17:52:18 crc kubenswrapper[4868]: I1201 17:52:18.045109 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-f1c8-account-create-update-p4sps"] Dec 01 17:52:18 crc kubenswrapper[4868]: I1201 17:52:18.056493 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-8qflm"] Dec 01 17:52:18 crc kubenswrapper[4868]: I1201 17:52:18.064862 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-f1c8-account-create-update-p4sps"] Dec 01 17:52:18 crc kubenswrapper[4868]: I1201 17:52:18.183344 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35b2d43b-53b4-4b3b-abce-5393c44d1db9" path="/var/lib/kubelet/pods/35b2d43b-53b4-4b3b-abce-5393c44d1db9/volumes" Dec 01 17:52:18 crc kubenswrapper[4868]: I1201 17:52:18.183994 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b47b287b-3cf1-4814-a47a-76bb02275bb8" path="/var/lib/kubelet/pods/b47b287b-3cf1-4814-a47a-76bb02275bb8/volumes" Dec 01 17:52:18 crc kubenswrapper[4868]: I1201 17:52:18.184616 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cab673b6-4d49-47e9-8b7b-54dabbae8e81" path="/var/lib/kubelet/pods/cab673b6-4d49-47e9-8b7b-54dabbae8e81/volumes" Dec 01 17:52:18 crc kubenswrapper[4868]: I1201 17:52:18.185156 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cba4291a-4880-41fa-81ad-7df9b4287511" path="/var/lib/kubelet/pods/cba4291a-4880-41fa-81ad-7df9b4287511/volumes" Dec 01 17:52:25 crc kubenswrapper[4868]: I1201 17:52:25.904620 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:52:25 crc kubenswrapper[4868]: I1201 17:52:25.905703 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:52:39 crc kubenswrapper[4868]: I1201 17:52:39.113466 4868 scope.go:117] "RemoveContainer" containerID="7cd4dc57fd9bd34ab783eabf2fc54393fd2ee133a1fb481773fc1cdf00c398e2" Dec 01 17:52:39 crc kubenswrapper[4868]: I1201 17:52:39.156584 4868 scope.go:117] "RemoveContainer" containerID="6067bb7b9f378fe293a88744787d62b6d33587fd676beec9c748c509edfe2da2" Dec 01 17:52:39 crc kubenswrapper[4868]: I1201 17:52:39.226522 4868 scope.go:117] "RemoveContainer" containerID="0be7ba16763099d2d5465f8dd9f7c6e29a71a14ca61ceac2525dd920b54b1dd3" Dec 01 17:52:39 crc kubenswrapper[4868]: I1201 17:52:39.276152 4868 scope.go:117] "RemoveContainer" containerID="f6e962410e6d04cbee7c7f78ac64f6ae6bc7acc8e4033205654f21e6009eaf91" Dec 01 17:52:39 crc kubenswrapper[4868]: I1201 17:52:39.330903 4868 scope.go:117] "RemoveContainer" containerID="aa77b0e41cb09a8344c67ce86e528c0d2af863c8cf6d415c25ce5334e8e03de7" Dec 01 17:52:39 crc kubenswrapper[4868]: I1201 17:52:39.385666 4868 scope.go:117] "RemoveContainer" containerID="71c96f022e567aaecd6b1e11c5cc6b53c845bd76468079b26cfcdfd944c11d11" Dec 01 17:52:39 crc kubenswrapper[4868]: I1201 17:52:39.430004 4868 scope.go:117] "RemoveContainer" containerID="093bbc91a298e5f48e40e82040e6a5b1956f89305df649897d67c8242fa90136" Dec 01 17:52:39 crc kubenswrapper[4868]: I1201 17:52:39.460499 4868 scope.go:117] "RemoveContainer" containerID="bd0ef08c1a95cec1b2f700e86f514dafb7ddd4f269f1b67a10250eac3896a6e4" Dec 01 17:52:46 crc kubenswrapper[4868]: I1201 17:52:46.050803 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-6z7mt"] Dec 01 17:52:46 crc kubenswrapper[4868]: I1201 17:52:46.059207 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-6z7mt"] Dec 01 17:52:46 crc kubenswrapper[4868]: I1201 17:52:46.182568 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2017518-df87-4ece-a8cf-4805edbf367b" path="/var/lib/kubelet/pods/a2017518-df87-4ece-a8cf-4805edbf367b/volumes" Dec 01 17:52:49 crc kubenswrapper[4868]: I1201 17:52:49.357051 4868 generic.go:334] "Generic (PLEG): container finished" podID="6e39e56d-e163-4e75-909f-410c72fde805" containerID="ef99afb01e648b4610541c68c101aa006cb8dd1b51cf387a4704a5e23a46f9c8" exitCode=0 Dec 01 17:52:49 crc kubenswrapper[4868]: I1201 17:52:49.357137 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" event={"ID":"6e39e56d-e163-4e75-909f-410c72fde805","Type":"ContainerDied","Data":"ef99afb01e648b4610541c68c101aa006cb8dd1b51cf387a4704a5e23a46f9c8"} Dec 01 17:52:50 crc kubenswrapper[4868]: I1201 17:52:50.745098 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" Dec 01 17:52:50 crc kubenswrapper[4868]: I1201 17:52:50.921673 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e39e56d-e163-4e75-909f-410c72fde805-ssh-key\") pod \"6e39e56d-e163-4e75-909f-410c72fde805\" (UID: \"6e39e56d-e163-4e75-909f-410c72fde805\") " Dec 01 17:52:50 crc kubenswrapper[4868]: I1201 17:52:50.922925 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b68pc\" (UniqueName: \"kubernetes.io/projected/6e39e56d-e163-4e75-909f-410c72fde805-kube-api-access-b68pc\") pod \"6e39e56d-e163-4e75-909f-410c72fde805\" (UID: \"6e39e56d-e163-4e75-909f-410c72fde805\") " Dec 01 17:52:50 crc kubenswrapper[4868]: I1201 17:52:50.923117 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e39e56d-e163-4e75-909f-410c72fde805-inventory\") pod \"6e39e56d-e163-4e75-909f-410c72fde805\" (UID: \"6e39e56d-e163-4e75-909f-410c72fde805\") " Dec 01 17:52:50 crc kubenswrapper[4868]: I1201 17:52:50.929228 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e39e56d-e163-4e75-909f-410c72fde805-kube-api-access-b68pc" (OuterVolumeSpecName: "kube-api-access-b68pc") pod "6e39e56d-e163-4e75-909f-410c72fde805" (UID: "6e39e56d-e163-4e75-909f-410c72fde805"). InnerVolumeSpecName "kube-api-access-b68pc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:52:50 crc kubenswrapper[4868]: I1201 17:52:50.953111 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e39e56d-e163-4e75-909f-410c72fde805-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6e39e56d-e163-4e75-909f-410c72fde805" (UID: "6e39e56d-e163-4e75-909f-410c72fde805"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:52:50 crc kubenswrapper[4868]: I1201 17:52:50.953859 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e39e56d-e163-4e75-909f-410c72fde805-inventory" (OuterVolumeSpecName: "inventory") pod "6e39e56d-e163-4e75-909f-410c72fde805" (UID: "6e39e56d-e163-4e75-909f-410c72fde805"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.025876 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e39e56d-e163-4e75-909f-410c72fde805-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.025911 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e39e56d-e163-4e75-909f-410c72fde805-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.025923 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b68pc\" (UniqueName: \"kubernetes.io/projected/6e39e56d-e163-4e75-909f-410c72fde805-kube-api-access-b68pc\") on node \"crc\" DevicePath \"\"" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.378831 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" event={"ID":"6e39e56d-e163-4e75-909f-410c72fde805","Type":"ContainerDied","Data":"462de88fbde64e2a3b81bdc38090df6fa08aa2448dfa85611db5bc1b91969538"} Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.379393 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="462de88fbde64e2a3b81bdc38090df6fa08aa2448dfa85611db5bc1b91969538" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.378866 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.482151 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg"] Dec 01 17:52:51 crc kubenswrapper[4868]: E1201 17:52:51.482557 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e39e56d-e163-4e75-909f-410c72fde805" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.482583 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e39e56d-e163-4e75-909f-410c72fde805" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 01 17:52:51 crc kubenswrapper[4868]: E1201 17:52:51.482613 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37079409-9062-477e-98a5-ca24163cef35" containerName="extract-utilities" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.482621 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="37079409-9062-477e-98a5-ca24163cef35" containerName="extract-utilities" Dec 01 17:52:51 crc kubenswrapper[4868]: E1201 17:52:51.482655 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37079409-9062-477e-98a5-ca24163cef35" containerName="registry-server" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.482662 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="37079409-9062-477e-98a5-ca24163cef35" containerName="registry-server" Dec 01 17:52:51 crc kubenswrapper[4868]: E1201 17:52:51.482681 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37079409-9062-477e-98a5-ca24163cef35" containerName="extract-content" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.482687 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="37079409-9062-477e-98a5-ca24163cef35" containerName="extract-content" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.482893 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="37079409-9062-477e-98a5-ca24163cef35" containerName="registry-server" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.482913 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e39e56d-e163-4e75-909f-410c72fde805" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.483531 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.486214 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.486598 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.486825 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.487707 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.505170 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg"] Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.640669 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8krlq\" (UniqueName: \"kubernetes.io/projected/fb542a8c-69c4-4a30-9a43-de20584d8e23-kube-api-access-8krlq\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg\" (UID: \"fb542a8c-69c4-4a30-9a43-de20584d8e23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.640780 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb542a8c-69c4-4a30-9a43-de20584d8e23-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg\" (UID: \"fb542a8c-69c4-4a30-9a43-de20584d8e23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.640840 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb542a8c-69c4-4a30-9a43-de20584d8e23-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg\" (UID: \"fb542a8c-69c4-4a30-9a43-de20584d8e23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.744124 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8krlq\" (UniqueName: \"kubernetes.io/projected/fb542a8c-69c4-4a30-9a43-de20584d8e23-kube-api-access-8krlq\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg\" (UID: \"fb542a8c-69c4-4a30-9a43-de20584d8e23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.744893 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb542a8c-69c4-4a30-9a43-de20584d8e23-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg\" (UID: \"fb542a8c-69c4-4a30-9a43-de20584d8e23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.746411 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb542a8c-69c4-4a30-9a43-de20584d8e23-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg\" (UID: \"fb542a8c-69c4-4a30-9a43-de20584d8e23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.751154 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb542a8c-69c4-4a30-9a43-de20584d8e23-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg\" (UID: \"fb542a8c-69c4-4a30-9a43-de20584d8e23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.754200 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb542a8c-69c4-4a30-9a43-de20584d8e23-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg\" (UID: \"fb542a8c-69c4-4a30-9a43-de20584d8e23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.761356 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8krlq\" (UniqueName: \"kubernetes.io/projected/fb542a8c-69c4-4a30-9a43-de20584d8e23-kube-api-access-8krlq\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg\" (UID: \"fb542a8c-69c4-4a30-9a43-de20584d8e23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" Dec 01 17:52:51 crc kubenswrapper[4868]: I1201 17:52:51.848913 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" Dec 01 17:52:52 crc kubenswrapper[4868]: I1201 17:52:52.465458 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg"] Dec 01 17:52:53 crc kubenswrapper[4868]: I1201 17:52:53.421369 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" event={"ID":"fb542a8c-69c4-4a30-9a43-de20584d8e23","Type":"ContainerStarted","Data":"3cf8ea02f4748d264f598a424b7acc4e8d1636b070102bcc389f584a2051e401"} Dec 01 17:52:53 crc kubenswrapper[4868]: I1201 17:52:53.422115 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" event={"ID":"fb542a8c-69c4-4a30-9a43-de20584d8e23","Type":"ContainerStarted","Data":"24c621ef7762b7e0d30ffef524762f8c4cf717f32db1ea99e131c1c805f3f7ab"} Dec 01 17:52:53 crc kubenswrapper[4868]: I1201 17:52:53.440363 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" podStartSLOduration=1.841092221 podStartE2EDuration="2.440342101s" podCreationTimestamp="2025-12-01 17:52:51 +0000 UTC" firstStartedPulling="2025-12-01 17:52:52.478526516 +0000 UTC m=+1644.849636947" lastFinishedPulling="2025-12-01 17:52:53.077776396 +0000 UTC m=+1645.448886827" observedRunningTime="2025-12-01 17:52:53.437864483 +0000 UTC m=+1645.808974904" watchObservedRunningTime="2025-12-01 17:52:53.440342101 +0000 UTC m=+1645.811452512" Dec 01 17:52:55 crc kubenswrapper[4868]: I1201 17:52:55.033182 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-hmdjz"] Dec 01 17:52:55 crc kubenswrapper[4868]: I1201 17:52:55.044702 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-6b6d-account-create-update-jnz42"] Dec 01 17:52:55 crc kubenswrapper[4868]: I1201 17:52:55.055671 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-hmdjz"] Dec 01 17:52:55 crc kubenswrapper[4868]: I1201 17:52:55.063373 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-fwvm2"] Dec 01 17:52:55 crc kubenswrapper[4868]: I1201 17:52:55.070667 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-fwvm2"] Dec 01 17:52:55 crc kubenswrapper[4868]: I1201 17:52:55.077414 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-6b6d-account-create-update-jnz42"] Dec 01 17:52:55 crc kubenswrapper[4868]: I1201 17:52:55.905208 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:52:55 crc kubenswrapper[4868]: I1201 17:52:55.905591 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:52:56 crc kubenswrapper[4868]: I1201 17:52:56.182205 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4a29c27-3526-46bf-a3d2-3aacde054e78" path="/var/lib/kubelet/pods/b4a29c27-3526-46bf-a3d2-3aacde054e78/volumes" Dec 01 17:52:56 crc kubenswrapper[4868]: I1201 17:52:56.182770 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db6357f7-cfa8-4f21-8b9f-223f828420c3" path="/var/lib/kubelet/pods/db6357f7-cfa8-4f21-8b9f-223f828420c3/volumes" Dec 01 17:52:56 crc kubenswrapper[4868]: I1201 17:52:56.183340 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efe2d244-cf00-49dc-9423-fe757ac195d3" path="/var/lib/kubelet/pods/efe2d244-cf00-49dc-9423-fe757ac195d3/volumes" Dec 01 17:52:59 crc kubenswrapper[4868]: I1201 17:52:59.037399 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-9zxn4"] Dec 01 17:52:59 crc kubenswrapper[4868]: I1201 17:52:59.052127 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d9a-account-create-update-2mbn6"] Dec 01 17:52:59 crc kubenswrapper[4868]: I1201 17:52:59.059237 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-af0b-account-create-update-wzxbn"] Dec 01 17:52:59 crc kubenswrapper[4868]: I1201 17:52:59.067368 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-af0b-account-create-update-wzxbn"] Dec 01 17:52:59 crc kubenswrapper[4868]: I1201 17:52:59.075473 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-9zxn4"] Dec 01 17:52:59 crc kubenswrapper[4868]: I1201 17:52:59.085561 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5d9a-account-create-update-2mbn6"] Dec 01 17:53:00 crc kubenswrapper[4868]: I1201 17:53:00.185972 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="641f63ea-e837-4644-99af-c83045d72395" path="/var/lib/kubelet/pods/641f63ea-e837-4644-99af-c83045d72395/volumes" Dec 01 17:53:00 crc kubenswrapper[4868]: I1201 17:53:00.187995 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d9c1c1c-859d-4b2a-9e6b-1a3d863df282" path="/var/lib/kubelet/pods/6d9c1c1c-859d-4b2a-9e6b-1a3d863df282/volumes" Dec 01 17:53:00 crc kubenswrapper[4868]: I1201 17:53:00.188616 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e6e56df-4807-42c7-a850-4369abb6e61e" path="/var/lib/kubelet/pods/8e6e56df-4807-42c7-a850-4369abb6e61e/volumes" Dec 01 17:53:03 crc kubenswrapper[4868]: I1201 17:53:03.043740 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-dnn68"] Dec 01 17:53:03 crc kubenswrapper[4868]: I1201 17:53:03.052449 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-dnn68"] Dec 01 17:53:04 crc kubenswrapper[4868]: I1201 17:53:04.187294 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b30342a-d0e7-4507-8968-05904e98ed9a" path="/var/lib/kubelet/pods/7b30342a-d0e7-4507-8968-05904e98ed9a/volumes" Dec 01 17:53:25 crc kubenswrapper[4868]: I1201 17:53:25.905268 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:53:25 crc kubenswrapper[4868]: I1201 17:53:25.906190 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:53:25 crc kubenswrapper[4868]: I1201 17:53:25.906266 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 17:53:25 crc kubenswrapper[4868]: I1201 17:53:25.907470 4868 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded"} pod="openshift-machine-config-operator/machine-config-daemon-njgpn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 17:53:25 crc kubenswrapper[4868]: I1201 17:53:25.907550 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" containerID="cri-o://5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" gracePeriod=600 Dec 01 17:53:26 crc kubenswrapper[4868]: E1201 17:53:26.049797 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:53:26 crc kubenswrapper[4868]: I1201 17:53:26.971156 4868 generic.go:334] "Generic (PLEG): container finished" podID="ec4e08bf-df56-4490-8e79-a4175c727405" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" exitCode=0 Dec 01 17:53:26 crc kubenswrapper[4868]: I1201 17:53:26.971237 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerDied","Data":"5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded"} Dec 01 17:53:26 crc kubenswrapper[4868]: I1201 17:53:26.971299 4868 scope.go:117] "RemoveContainer" containerID="998e9f9a5d44e576ae1e608687cc6cddbcaf1995065246b296eae79bc764efdb" Dec 01 17:53:26 crc kubenswrapper[4868]: I1201 17:53:26.972600 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:53:26 crc kubenswrapper[4868]: E1201 17:53:26.973118 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:53:35 crc kubenswrapper[4868]: I1201 17:53:35.037279 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-fbgdl"] Dec 01 17:53:35 crc kubenswrapper[4868]: I1201 17:53:35.046211 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-fbgdl"] Dec 01 17:53:36 crc kubenswrapper[4868]: I1201 17:53:36.193475 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27e82e73-8d20-4b1f-9514-11aad65f4331" path="/var/lib/kubelet/pods/27e82e73-8d20-4b1f-9514-11aad65f4331/volumes" Dec 01 17:53:38 crc kubenswrapper[4868]: I1201 17:53:38.182435 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:53:38 crc kubenswrapper[4868]: E1201 17:53:38.183008 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:53:39 crc kubenswrapper[4868]: I1201 17:53:39.700105 4868 scope.go:117] "RemoveContainer" containerID="3937365a949c713047e17c30856b3a6190c4b98afdcaf20c8a433ab9851f6e3d" Dec 01 17:53:39 crc kubenswrapper[4868]: I1201 17:53:39.747551 4868 scope.go:117] "RemoveContainer" containerID="c70a918b6b5630a265c0fdfb8dbc21f7c478a51f339e0f48254f56d912e7a052" Dec 01 17:53:39 crc kubenswrapper[4868]: I1201 17:53:39.788310 4868 scope.go:117] "RemoveContainer" containerID="58948cf462aa3d2b0ee2ea3d790a38977a1092272155fc8fcfaa5be21250a6d8" Dec 01 17:53:39 crc kubenswrapper[4868]: I1201 17:53:39.857652 4868 scope.go:117] "RemoveContainer" containerID="ff5c1ea23acc346cf2d00fa44c161aa13ee4092a071208bfce781277cc369d92" Dec 01 17:53:39 crc kubenswrapper[4868]: I1201 17:53:39.879937 4868 scope.go:117] "RemoveContainer" containerID="91d07288ef7ac072439586ff0d732c58771e82e82f3785f5f153b621bfff759b" Dec 01 17:53:39 crc kubenswrapper[4868]: I1201 17:53:39.930547 4868 scope.go:117] "RemoveContainer" containerID="bb683e475676090d0da843b669bb2d444dd1c46715f86f31e45d7d5d91c5e3ba" Dec 01 17:53:39 crc kubenswrapper[4868]: I1201 17:53:39.972710 4868 scope.go:117] "RemoveContainer" containerID="cd7e61c92876982a7e924490b9dd7414a5c021c25911071f3ac28d7684e0740e" Dec 01 17:53:40 crc kubenswrapper[4868]: I1201 17:53:40.001504 4868 scope.go:117] "RemoveContainer" containerID="45963812b655d654c05a615cbeb3bddfff82bba9799828eea2cbdd413b56f694" Dec 01 17:53:40 crc kubenswrapper[4868]: I1201 17:53:40.045464 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-mrzhs"] Dec 01 17:53:40 crc kubenswrapper[4868]: I1201 17:53:40.056706 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-mrzhs"] Dec 01 17:53:40 crc kubenswrapper[4868]: I1201 17:53:40.081586 4868 scope.go:117] "RemoveContainer" containerID="a3eddc1407dda99e2d67d97f23e6008b027a9ac998e0680735a4a33f4d37628c" Dec 01 17:53:40 crc kubenswrapper[4868]: I1201 17:53:40.192162 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8a7a195-5f45-470b-acfc-db3d4b69ca12" path="/var/lib/kubelet/pods/f8a7a195-5f45-470b-acfc-db3d4b69ca12/volumes" Dec 01 17:53:50 crc kubenswrapper[4868]: I1201 17:53:50.040854 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-8stjq"] Dec 01 17:53:50 crc kubenswrapper[4868]: I1201 17:53:50.057177 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-8stjq"] Dec 01 17:53:50 crc kubenswrapper[4868]: I1201 17:53:50.070103 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-hf8rq"] Dec 01 17:53:50 crc kubenswrapper[4868]: I1201 17:53:50.081463 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-hf8rq"] Dec 01 17:53:50 crc kubenswrapper[4868]: I1201 17:53:50.173192 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:53:50 crc kubenswrapper[4868]: E1201 17:53:50.173774 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:53:50 crc kubenswrapper[4868]: I1201 17:53:50.182759 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be92b184-744a-4da9-be7f-2441b7d1452d" path="/var/lib/kubelet/pods/be92b184-744a-4da9-be7f-2441b7d1452d/volumes" Dec 01 17:53:50 crc kubenswrapper[4868]: I1201 17:53:50.183664 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed070cb2-83dc-4f53-97ed-0ffff0baab82" path="/var/lib/kubelet/pods/ed070cb2-83dc-4f53-97ed-0ffff0baab82/volumes" Dec 01 17:54:01 crc kubenswrapper[4868]: I1201 17:54:01.371838 4868 generic.go:334] "Generic (PLEG): container finished" podID="fb542a8c-69c4-4a30-9a43-de20584d8e23" containerID="3cf8ea02f4748d264f598a424b7acc4e8d1636b070102bcc389f584a2051e401" exitCode=0 Dec 01 17:54:01 crc kubenswrapper[4868]: I1201 17:54:01.371953 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" event={"ID":"fb542a8c-69c4-4a30-9a43-de20584d8e23","Type":"ContainerDied","Data":"3cf8ea02f4748d264f598a424b7acc4e8d1636b070102bcc389f584a2051e401"} Dec 01 17:54:02 crc kubenswrapper[4868]: I1201 17:54:02.030617 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-vbxmg"] Dec 01 17:54:02 crc kubenswrapper[4868]: I1201 17:54:02.041519 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-vbxmg"] Dec 01 17:54:02 crc kubenswrapper[4868]: I1201 17:54:02.182433 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88c4f381-e8fa-474a-8f31-d0e2745ab83d" path="/var/lib/kubelet/pods/88c4f381-e8fa-474a-8f31-d0e2745ab83d/volumes" Dec 01 17:54:02 crc kubenswrapper[4868]: I1201 17:54:02.793135 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" Dec 01 17:54:02 crc kubenswrapper[4868]: I1201 17:54:02.881829 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb542a8c-69c4-4a30-9a43-de20584d8e23-ssh-key\") pod \"fb542a8c-69c4-4a30-9a43-de20584d8e23\" (UID: \"fb542a8c-69c4-4a30-9a43-de20584d8e23\") " Dec 01 17:54:02 crc kubenswrapper[4868]: I1201 17:54:02.882288 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8krlq\" (UniqueName: \"kubernetes.io/projected/fb542a8c-69c4-4a30-9a43-de20584d8e23-kube-api-access-8krlq\") pod \"fb542a8c-69c4-4a30-9a43-de20584d8e23\" (UID: \"fb542a8c-69c4-4a30-9a43-de20584d8e23\") " Dec 01 17:54:02 crc kubenswrapper[4868]: I1201 17:54:02.882545 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb542a8c-69c4-4a30-9a43-de20584d8e23-inventory\") pod \"fb542a8c-69c4-4a30-9a43-de20584d8e23\" (UID: \"fb542a8c-69c4-4a30-9a43-de20584d8e23\") " Dec 01 17:54:02 crc kubenswrapper[4868]: I1201 17:54:02.889248 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb542a8c-69c4-4a30-9a43-de20584d8e23-kube-api-access-8krlq" (OuterVolumeSpecName: "kube-api-access-8krlq") pod "fb542a8c-69c4-4a30-9a43-de20584d8e23" (UID: "fb542a8c-69c4-4a30-9a43-de20584d8e23"). InnerVolumeSpecName "kube-api-access-8krlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:54:02 crc kubenswrapper[4868]: I1201 17:54:02.911100 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb542a8c-69c4-4a30-9a43-de20584d8e23-inventory" (OuterVolumeSpecName: "inventory") pod "fb542a8c-69c4-4a30-9a43-de20584d8e23" (UID: "fb542a8c-69c4-4a30-9a43-de20584d8e23"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:54:02 crc kubenswrapper[4868]: I1201 17:54:02.911800 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb542a8c-69c4-4a30-9a43-de20584d8e23-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fb542a8c-69c4-4a30-9a43-de20584d8e23" (UID: "fb542a8c-69c4-4a30-9a43-de20584d8e23"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:54:02 crc kubenswrapper[4868]: I1201 17:54:02.984294 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb542a8c-69c4-4a30-9a43-de20584d8e23-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:54:02 crc kubenswrapper[4868]: I1201 17:54:02.984341 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb542a8c-69c4-4a30-9a43-de20584d8e23-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:54:02 crc kubenswrapper[4868]: I1201 17:54:02.984354 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8krlq\" (UniqueName: \"kubernetes.io/projected/fb542a8c-69c4-4a30-9a43-de20584d8e23-kube-api-access-8krlq\") on node \"crc\" DevicePath \"\"" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.388494 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.388426 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg" event={"ID":"fb542a8c-69c4-4a30-9a43-de20584d8e23","Type":"ContainerDied","Data":"24c621ef7762b7e0d30ffef524762f8c4cf717f32db1ea99e131c1c805f3f7ab"} Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.388566 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24c621ef7762b7e0d30ffef524762f8c4cf717f32db1ea99e131c1c805f3f7ab" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.484994 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68"] Dec 01 17:54:03 crc kubenswrapper[4868]: E1201 17:54:03.485411 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb542a8c-69c4-4a30-9a43-de20584d8e23" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.485425 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb542a8c-69c4-4a30-9a43-de20584d8e23" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.485633 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb542a8c-69c4-4a30-9a43-de20584d8e23" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.486549 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.491898 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.492108 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.492310 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.492438 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.515672 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68"] Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.596990 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f1d81003-13ab-4833-9b01-3b4c6f8876ef-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-v6t68\" (UID: \"f1d81003-13ab-4833-9b01-3b4c6f8876ef\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.597385 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5fq8\" (UniqueName: \"kubernetes.io/projected/f1d81003-13ab-4833-9b01-3b4c6f8876ef-kube-api-access-d5fq8\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-v6t68\" (UID: \"f1d81003-13ab-4833-9b01-3b4c6f8876ef\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.597421 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f1d81003-13ab-4833-9b01-3b4c6f8876ef-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-v6t68\" (UID: \"f1d81003-13ab-4833-9b01-3b4c6f8876ef\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.698878 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f1d81003-13ab-4833-9b01-3b4c6f8876ef-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-v6t68\" (UID: \"f1d81003-13ab-4833-9b01-3b4c6f8876ef\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.698963 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5fq8\" (UniqueName: \"kubernetes.io/projected/f1d81003-13ab-4833-9b01-3b4c6f8876ef-kube-api-access-d5fq8\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-v6t68\" (UID: \"f1d81003-13ab-4833-9b01-3b4c6f8876ef\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.699004 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f1d81003-13ab-4833-9b01-3b4c6f8876ef-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-v6t68\" (UID: \"f1d81003-13ab-4833-9b01-3b4c6f8876ef\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.704182 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f1d81003-13ab-4833-9b01-3b4c6f8876ef-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-v6t68\" (UID: \"f1d81003-13ab-4833-9b01-3b4c6f8876ef\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.713011 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f1d81003-13ab-4833-9b01-3b4c6f8876ef-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-v6t68\" (UID: \"f1d81003-13ab-4833-9b01-3b4c6f8876ef\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.719590 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5fq8\" (UniqueName: \"kubernetes.io/projected/f1d81003-13ab-4833-9b01-3b4c6f8876ef-kube-api-access-d5fq8\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-v6t68\" (UID: \"f1d81003-13ab-4833-9b01-3b4c6f8876ef\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" Dec 01 17:54:03 crc kubenswrapper[4868]: I1201 17:54:03.817587 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" Dec 01 17:54:04 crc kubenswrapper[4868]: I1201 17:54:04.312291 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68"] Dec 01 17:54:04 crc kubenswrapper[4868]: I1201 17:54:04.399682 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" event={"ID":"f1d81003-13ab-4833-9b01-3b4c6f8876ef","Type":"ContainerStarted","Data":"38486398dfabb5316a25154f60f1c9b7a535f86c3ce7be8242bd59426a08b117"} Dec 01 17:54:05 crc kubenswrapper[4868]: I1201 17:54:05.172132 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:54:05 crc kubenswrapper[4868]: E1201 17:54:05.172731 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:54:06 crc kubenswrapper[4868]: I1201 17:54:06.419654 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" event={"ID":"f1d81003-13ab-4833-9b01-3b4c6f8876ef","Type":"ContainerStarted","Data":"f98fc5efbdcd52a3ef710758a48e8767e95925ff1c5d981774bc010313192972"} Dec 01 17:54:06 crc kubenswrapper[4868]: I1201 17:54:06.445744 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" podStartSLOduration=2.529785799 podStartE2EDuration="3.445710129s" podCreationTimestamp="2025-12-01 17:54:03 +0000 UTC" firstStartedPulling="2025-12-01 17:54:04.318358171 +0000 UTC m=+1716.689468582" lastFinishedPulling="2025-12-01 17:54:05.234282461 +0000 UTC m=+1717.605392912" observedRunningTime="2025-12-01 17:54:06.438352529 +0000 UTC m=+1718.809462950" watchObservedRunningTime="2025-12-01 17:54:06.445710129 +0000 UTC m=+1718.816820540" Dec 01 17:54:10 crc kubenswrapper[4868]: I1201 17:54:10.462812 4868 generic.go:334] "Generic (PLEG): container finished" podID="f1d81003-13ab-4833-9b01-3b4c6f8876ef" containerID="f98fc5efbdcd52a3ef710758a48e8767e95925ff1c5d981774bc010313192972" exitCode=0 Dec 01 17:54:10 crc kubenswrapper[4868]: I1201 17:54:10.463692 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" event={"ID":"f1d81003-13ab-4833-9b01-3b4c6f8876ef","Type":"ContainerDied","Data":"f98fc5efbdcd52a3ef710758a48e8767e95925ff1c5d981774bc010313192972"} Dec 01 17:54:11 crc kubenswrapper[4868]: I1201 17:54:11.954412 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.052637 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9"] Dec 01 17:54:12 crc kubenswrapper[4868]: E1201 17:54:12.053888 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1d81003-13ab-4833-9b01-3b4c6f8876ef" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.053918 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1d81003-13ab-4833-9b01-3b4c6f8876ef" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.054318 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1d81003-13ab-4833-9b01-3b4c6f8876ef" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.055431 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.069493 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9"] Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.114720 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5fq8\" (UniqueName: \"kubernetes.io/projected/f1d81003-13ab-4833-9b01-3b4c6f8876ef-kube-api-access-d5fq8\") pod \"f1d81003-13ab-4833-9b01-3b4c6f8876ef\" (UID: \"f1d81003-13ab-4833-9b01-3b4c6f8876ef\") " Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.115189 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f1d81003-13ab-4833-9b01-3b4c6f8876ef-inventory\") pod \"f1d81003-13ab-4833-9b01-3b4c6f8876ef\" (UID: \"f1d81003-13ab-4833-9b01-3b4c6f8876ef\") " Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.115321 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f1d81003-13ab-4833-9b01-3b4c6f8876ef-ssh-key\") pod \"f1d81003-13ab-4833-9b01-3b4c6f8876ef\" (UID: \"f1d81003-13ab-4833-9b01-3b4c6f8876ef\") " Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.122204 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1d81003-13ab-4833-9b01-3b4c6f8876ef-kube-api-access-d5fq8" (OuterVolumeSpecName: "kube-api-access-d5fq8") pod "f1d81003-13ab-4833-9b01-3b4c6f8876ef" (UID: "f1d81003-13ab-4833-9b01-3b4c6f8876ef"). InnerVolumeSpecName "kube-api-access-d5fq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.144496 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1d81003-13ab-4833-9b01-3b4c6f8876ef-inventory" (OuterVolumeSpecName: "inventory") pod "f1d81003-13ab-4833-9b01-3b4c6f8876ef" (UID: "f1d81003-13ab-4833-9b01-3b4c6f8876ef"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.144886 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1d81003-13ab-4833-9b01-3b4c6f8876ef-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f1d81003-13ab-4833-9b01-3b4c6f8876ef" (UID: "f1d81003-13ab-4833-9b01-3b4c6f8876ef"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.218097 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5x8p9\" (UID: \"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.218279 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtd64\" (UniqueName: \"kubernetes.io/projected/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-kube-api-access-dtd64\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5x8p9\" (UID: \"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.218463 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5x8p9\" (UID: \"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.218605 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5fq8\" (UniqueName: \"kubernetes.io/projected/f1d81003-13ab-4833-9b01-3b4c6f8876ef-kube-api-access-d5fq8\") on node \"crc\" DevicePath \"\"" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.218631 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f1d81003-13ab-4833-9b01-3b4c6f8876ef-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.218649 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f1d81003-13ab-4833-9b01-3b4c6f8876ef-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.321434 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5x8p9\" (UID: \"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.322662 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtd64\" (UniqueName: \"kubernetes.io/projected/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-kube-api-access-dtd64\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5x8p9\" (UID: \"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.323114 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5x8p9\" (UID: \"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.326599 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5x8p9\" (UID: \"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.327367 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5x8p9\" (UID: \"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.347446 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtd64\" (UniqueName: \"kubernetes.io/projected/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-kube-api-access-dtd64\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5x8p9\" (UID: \"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.386769 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.487312 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" event={"ID":"f1d81003-13ab-4833-9b01-3b4c6f8876ef","Type":"ContainerDied","Data":"38486398dfabb5316a25154f60f1c9b7a535f86c3ce7be8242bd59426a08b117"} Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.487361 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38486398dfabb5316a25154f60f1c9b7a535f86c3ce7be8242bd59426a08b117" Dec 01 17:54:12 crc kubenswrapper[4868]: I1201 17:54:12.487377 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-v6t68" Dec 01 17:54:13 crc kubenswrapper[4868]: I1201 17:54:13.040680 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9"] Dec 01 17:54:13 crc kubenswrapper[4868]: I1201 17:54:13.531064 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" event={"ID":"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c","Type":"ContainerStarted","Data":"aad2383c62650c25d211f29e8eacdc84816b07efb89a54f572f8f01c503b6819"} Dec 01 17:54:14 crc kubenswrapper[4868]: I1201 17:54:14.544232 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" event={"ID":"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c","Type":"ContainerStarted","Data":"d056f3e035b30ad9d81a3f892e2a08da244b3f5662a739d511e8d9b4cc554ff8"} Dec 01 17:54:14 crc kubenswrapper[4868]: I1201 17:54:14.565840 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" podStartSLOduration=1.7360083720000001 podStartE2EDuration="2.5658152s" podCreationTimestamp="2025-12-01 17:54:12 +0000 UTC" firstStartedPulling="2025-12-01 17:54:13.041820442 +0000 UTC m=+1725.412930853" lastFinishedPulling="2025-12-01 17:54:13.87162727 +0000 UTC m=+1726.242737681" observedRunningTime="2025-12-01 17:54:14.560324141 +0000 UTC m=+1726.931434592" watchObservedRunningTime="2025-12-01 17:54:14.5658152 +0000 UTC m=+1726.936925611" Dec 01 17:54:16 crc kubenswrapper[4868]: I1201 17:54:16.172934 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:54:16 crc kubenswrapper[4868]: E1201 17:54:16.173937 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:54:31 crc kubenswrapper[4868]: I1201 17:54:31.172131 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:54:31 crc kubenswrapper[4868]: E1201 17:54:31.172902 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:54:40 crc kubenswrapper[4868]: I1201 17:54:40.044427 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-xgtmd"] Dec 01 17:54:40 crc kubenswrapper[4868]: I1201 17:54:40.058919 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-xgtmd"] Dec 01 17:54:40 crc kubenswrapper[4868]: I1201 17:54:40.184925 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8af7d-f71e-42be-ab59-237f2d505b1f" path="/var/lib/kubelet/pods/31d8af7d-f71e-42be-ab59-237f2d505b1f/volumes" Dec 01 17:54:40 crc kubenswrapper[4868]: I1201 17:54:40.284156 4868 scope.go:117] "RemoveContainer" containerID="9affd763fff7f399f265876d93260f2dc47a8b6ea25026ae7646d19a9d469498" Dec 01 17:54:40 crc kubenswrapper[4868]: I1201 17:54:40.331174 4868 scope.go:117] "RemoveContainer" containerID="517fb49a344e36a158dadd142d7cbf3c247e78907f8947945cade0df6d54b42b" Dec 01 17:54:40 crc kubenswrapper[4868]: I1201 17:54:40.394808 4868 scope.go:117] "RemoveContainer" containerID="8b0f1c867c810bfa762034c066053aa1e68fc4f3e7063c956aec64f2ef3f0c20" Dec 01 17:54:40 crc kubenswrapper[4868]: I1201 17:54:40.443378 4868 scope.go:117] "RemoveContainer" containerID="1a5f10048b30cc2ad962b3f2b9c71894e13a17c8ca71dd957944f89f55c28409" Dec 01 17:54:40 crc kubenswrapper[4868]: I1201 17:54:40.518637 4868 scope.go:117] "RemoveContainer" containerID="f779a6a1dfcf0682564edb209202e168825be2e615be4ed8dbcc17bc19bd18ca" Dec 01 17:54:41 crc kubenswrapper[4868]: I1201 17:54:41.032046 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-3252-account-create-update-jrq2v"] Dec 01 17:54:41 crc kubenswrapper[4868]: I1201 17:54:41.038562 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-3252-account-create-update-jrq2v"] Dec 01 17:54:42 crc kubenswrapper[4868]: I1201 17:54:42.042250 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-bc8a-account-create-update-ht62v"] Dec 01 17:54:42 crc kubenswrapper[4868]: I1201 17:54:42.060869 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-xfwp6"] Dec 01 17:54:42 crc kubenswrapper[4868]: I1201 17:54:42.068606 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-0c2b-account-create-update-gqzxq"] Dec 01 17:54:42 crc kubenswrapper[4868]: I1201 17:54:42.075984 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-wh5cc"] Dec 01 17:54:42 crc kubenswrapper[4868]: I1201 17:54:42.082604 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-bc8a-account-create-update-ht62v"] Dec 01 17:54:42 crc kubenswrapper[4868]: I1201 17:54:42.089504 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-xfwp6"] Dec 01 17:54:42 crc kubenswrapper[4868]: I1201 17:54:42.098409 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-wh5cc"] Dec 01 17:54:42 crc kubenswrapper[4868]: I1201 17:54:42.107196 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-0c2b-account-create-update-gqzxq"] Dec 01 17:54:42 crc kubenswrapper[4868]: I1201 17:54:42.182763 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="288e3235-6690-4c67-9e4f-2735be76d249" path="/var/lib/kubelet/pods/288e3235-6690-4c67-9e4f-2735be76d249/volumes" Dec 01 17:54:42 crc kubenswrapper[4868]: I1201 17:54:42.183390 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="441fc0fe-f3fe-4e94-abd8-0d126b289042" path="/var/lib/kubelet/pods/441fc0fe-f3fe-4e94-abd8-0d126b289042/volumes" Dec 01 17:54:42 crc kubenswrapper[4868]: I1201 17:54:42.184022 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a996536c-2c87-4c80-90d6-c6758c109af1" path="/var/lib/kubelet/pods/a996536c-2c87-4c80-90d6-c6758c109af1/volumes" Dec 01 17:54:42 crc kubenswrapper[4868]: I1201 17:54:42.184723 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b801bf8a-a39f-4537-aa83-06e958c31e02" path="/var/lib/kubelet/pods/b801bf8a-a39f-4537-aa83-06e958c31e02/volumes" Dec 01 17:54:42 crc kubenswrapper[4868]: I1201 17:54:42.185765 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5951a13-5101-4956-bcf6-345fab68ea75" path="/var/lib/kubelet/pods/e5951a13-5101-4956-bcf6-345fab68ea75/volumes" Dec 01 17:54:45 crc kubenswrapper[4868]: I1201 17:54:45.171689 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:54:45 crc kubenswrapper[4868]: E1201 17:54:45.172269 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:54:49 crc kubenswrapper[4868]: I1201 17:54:49.900245 4868 generic.go:334] "Generic (PLEG): container finished" podID="23a40d7f-6024-42d5-9f5f-cbdbc2c8740c" containerID="d056f3e035b30ad9d81a3f892e2a08da244b3f5662a739d511e8d9b4cc554ff8" exitCode=0 Dec 01 17:54:49 crc kubenswrapper[4868]: I1201 17:54:49.900312 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" event={"ID":"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c","Type":"ContainerDied","Data":"d056f3e035b30ad9d81a3f892e2a08da244b3f5662a739d511e8d9b4cc554ff8"} Dec 01 17:54:51 crc kubenswrapper[4868]: I1201 17:54:51.374345 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" Dec 01 17:54:51 crc kubenswrapper[4868]: I1201 17:54:51.495927 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-ssh-key\") pod \"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c\" (UID: \"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c\") " Dec 01 17:54:51 crc kubenswrapper[4868]: I1201 17:54:51.496026 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-inventory\") pod \"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c\" (UID: \"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c\") " Dec 01 17:54:51 crc kubenswrapper[4868]: I1201 17:54:51.496132 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtd64\" (UniqueName: \"kubernetes.io/projected/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-kube-api-access-dtd64\") pod \"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c\" (UID: \"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c\") " Dec 01 17:54:51 crc kubenswrapper[4868]: I1201 17:54:51.501108 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-kube-api-access-dtd64" (OuterVolumeSpecName: "kube-api-access-dtd64") pod "23a40d7f-6024-42d5-9f5f-cbdbc2c8740c" (UID: "23a40d7f-6024-42d5-9f5f-cbdbc2c8740c"). InnerVolumeSpecName "kube-api-access-dtd64". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:54:51 crc kubenswrapper[4868]: I1201 17:54:51.521366 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-inventory" (OuterVolumeSpecName: "inventory") pod "23a40d7f-6024-42d5-9f5f-cbdbc2c8740c" (UID: "23a40d7f-6024-42d5-9f5f-cbdbc2c8740c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:54:51 crc kubenswrapper[4868]: I1201 17:54:51.523454 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "23a40d7f-6024-42d5-9f5f-cbdbc2c8740c" (UID: "23a40d7f-6024-42d5-9f5f-cbdbc2c8740c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:54:51 crc kubenswrapper[4868]: I1201 17:54:51.598254 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:54:51 crc kubenswrapper[4868]: I1201 17:54:51.598295 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:54:51 crc kubenswrapper[4868]: I1201 17:54:51.598316 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtd64\" (UniqueName: \"kubernetes.io/projected/23a40d7f-6024-42d5-9f5f-cbdbc2c8740c-kube-api-access-dtd64\") on node \"crc\" DevicePath \"\"" Dec 01 17:54:51 crc kubenswrapper[4868]: I1201 17:54:51.926068 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" event={"ID":"23a40d7f-6024-42d5-9f5f-cbdbc2c8740c","Type":"ContainerDied","Data":"aad2383c62650c25d211f29e8eacdc84816b07efb89a54f572f8f01c503b6819"} Dec 01 17:54:51 crc kubenswrapper[4868]: I1201 17:54:51.926115 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aad2383c62650c25d211f29e8eacdc84816b07efb89a54f572f8f01c503b6819" Dec 01 17:54:51 crc kubenswrapper[4868]: I1201 17:54:51.926159 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5x8p9" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.010413 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8"] Dec 01 17:54:52 crc kubenswrapper[4868]: E1201 17:54:52.011120 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23a40d7f-6024-42d5-9f5f-cbdbc2c8740c" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.011139 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="23a40d7f-6024-42d5-9f5f-cbdbc2c8740c" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.011340 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="23a40d7f-6024-42d5-9f5f-cbdbc2c8740c" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.011997 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.017147 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.017163 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.019209 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.019444 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.026156 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8"] Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.106762 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rvcc\" (UniqueName: \"kubernetes.io/projected/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-kube-api-access-2rvcc\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8\" (UID: \"0d17c1fe-ab49-4510-a2cc-ed17ae06abad\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.106824 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8\" (UID: \"0d17c1fe-ab49-4510-a2cc-ed17ae06abad\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.106976 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8\" (UID: \"0d17c1fe-ab49-4510-a2cc-ed17ae06abad\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.208805 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rvcc\" (UniqueName: \"kubernetes.io/projected/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-kube-api-access-2rvcc\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8\" (UID: \"0d17c1fe-ab49-4510-a2cc-ed17ae06abad\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.208889 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8\" (UID: \"0d17c1fe-ab49-4510-a2cc-ed17ae06abad\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.208953 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8\" (UID: \"0d17c1fe-ab49-4510-a2cc-ed17ae06abad\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.214467 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8\" (UID: \"0d17c1fe-ab49-4510-a2cc-ed17ae06abad\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.215717 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8\" (UID: \"0d17c1fe-ab49-4510-a2cc-ed17ae06abad\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.231065 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rvcc\" (UniqueName: \"kubernetes.io/projected/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-kube-api-access-2rvcc\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8\" (UID: \"0d17c1fe-ab49-4510-a2cc-ed17ae06abad\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.328504 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.878789 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8"] Dec 01 17:54:52 crc kubenswrapper[4868]: I1201 17:54:52.938711 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" event={"ID":"0d17c1fe-ab49-4510-a2cc-ed17ae06abad","Type":"ContainerStarted","Data":"39d7649e3458017257d401590feb0fd4a71a6c58f31dd990ebed233b6a289d77"} Dec 01 17:54:53 crc kubenswrapper[4868]: I1201 17:54:53.954212 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" event={"ID":"0d17c1fe-ab49-4510-a2cc-ed17ae06abad","Type":"ContainerStarted","Data":"aae1575eb0879b2fa52b9b85dd8c80f16ae71641f5376eecf8c759537b6b9aab"} Dec 01 17:54:53 crc kubenswrapper[4868]: I1201 17:54:53.975212 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" podStartSLOduration=2.285666737 podStartE2EDuration="2.975194371s" podCreationTimestamp="2025-12-01 17:54:51 +0000 UTC" firstStartedPulling="2025-12-01 17:54:52.893190581 +0000 UTC m=+1765.264300992" lastFinishedPulling="2025-12-01 17:54:53.582718205 +0000 UTC m=+1765.953828626" observedRunningTime="2025-12-01 17:54:53.968776367 +0000 UTC m=+1766.339886778" watchObservedRunningTime="2025-12-01 17:54:53.975194371 +0000 UTC m=+1766.346304782" Dec 01 17:54:59 crc kubenswrapper[4868]: I1201 17:54:59.171884 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:54:59 crc kubenswrapper[4868]: E1201 17:54:59.172728 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:55:12 crc kubenswrapper[4868]: I1201 17:55:12.048983 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k7gsb"] Dec 01 17:55:12 crc kubenswrapper[4868]: I1201 17:55:12.057321 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k7gsb"] Dec 01 17:55:12 crc kubenswrapper[4868]: I1201 17:55:12.172669 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:55:12 crc kubenswrapper[4868]: E1201 17:55:12.173066 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:55:12 crc kubenswrapper[4868]: I1201 17:55:12.185793 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0f4d962-9aa7-4ae9-836c-cdce14d53159" path="/var/lib/kubelet/pods/d0f4d962-9aa7-4ae9-836c-cdce14d53159/volumes" Dec 01 17:55:23 crc kubenswrapper[4868]: I1201 17:55:23.172593 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:55:23 crc kubenswrapper[4868]: E1201 17:55:23.173671 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:55:34 crc kubenswrapper[4868]: I1201 17:55:34.038248 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-2tbrq"] Dec 01 17:55:34 crc kubenswrapper[4868]: I1201 17:55:34.045290 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-2tbrq"] Dec 01 17:55:34 crc kubenswrapper[4868]: I1201 17:55:34.186341 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb439a97-8909-414a-a4cf-22e94e909356" path="/var/lib/kubelet/pods/fb439a97-8909-414a-a4cf-22e94e909356/volumes" Dec 01 17:55:35 crc kubenswrapper[4868]: I1201 17:55:35.041154 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5pwdf"] Dec 01 17:55:35 crc kubenswrapper[4868]: I1201 17:55:35.051406 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5pwdf"] Dec 01 17:55:36 crc kubenswrapper[4868]: I1201 17:55:36.193089 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="897da32f-2dbb-416b-b58e-a4d59040c89c" path="/var/lib/kubelet/pods/897da32f-2dbb-416b-b58e-a4d59040c89c/volumes" Dec 01 17:55:37 crc kubenswrapper[4868]: I1201 17:55:37.173524 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:55:37 crc kubenswrapper[4868]: E1201 17:55:37.175712 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:55:40 crc kubenswrapper[4868]: I1201 17:55:40.661545 4868 scope.go:117] "RemoveContainer" containerID="adeebb3b056bc9d1d5a5af8450952e30d08eff9e76732c29baa0606725dae518" Dec 01 17:55:40 crc kubenswrapper[4868]: I1201 17:55:40.715518 4868 scope.go:117] "RemoveContainer" containerID="e2237ad0b6d8c1771316a95bd844ac047784f7c1e7107f6a7b93de8775347385" Dec 01 17:55:40 crc kubenswrapper[4868]: I1201 17:55:40.781281 4868 scope.go:117] "RemoveContainer" containerID="80b8fe19cab38a4094b4a12f3720205d98255899f44e9c3b71dc6387a57e148d" Dec 01 17:55:40 crc kubenswrapper[4868]: I1201 17:55:40.806176 4868 scope.go:117] "RemoveContainer" containerID="7d357e0acb6033a19187106e62ec3aa8e00109ee4c4372d7d5c56a01923a0553" Dec 01 17:55:40 crc kubenswrapper[4868]: I1201 17:55:40.856609 4868 scope.go:117] "RemoveContainer" containerID="418d436c4bbfe6f81c8b65515afb9a59fc9bd4230c8346a97475719320ad3d9a" Dec 01 17:55:40 crc kubenswrapper[4868]: I1201 17:55:40.909011 4868 scope.go:117] "RemoveContainer" containerID="0878faec332107965aaf8ad00d9fdf2258d16edd2098a54733a1bc3c889c111f" Dec 01 17:55:40 crc kubenswrapper[4868]: I1201 17:55:40.949812 4868 scope.go:117] "RemoveContainer" containerID="17e85ef8e3bd3ce5537753a9525c2ea00ed57ec7e412ffc6e943fddc8d20a79e" Dec 01 17:55:40 crc kubenswrapper[4868]: I1201 17:55:40.971765 4868 scope.go:117] "RemoveContainer" containerID="84f8dc40aac9df502e5e1ba67b84436cd7f524979e2974092d7a86954bb89e34" Dec 01 17:55:41 crc kubenswrapper[4868]: I1201 17:55:41.417050 4868 generic.go:334] "Generic (PLEG): container finished" podID="0d17c1fe-ab49-4510-a2cc-ed17ae06abad" containerID="aae1575eb0879b2fa52b9b85dd8c80f16ae71641f5376eecf8c759537b6b9aab" exitCode=0 Dec 01 17:55:41 crc kubenswrapper[4868]: I1201 17:55:41.417090 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" event={"ID":"0d17c1fe-ab49-4510-a2cc-ed17ae06abad","Type":"ContainerDied","Data":"aae1575eb0879b2fa52b9b85dd8c80f16ae71641f5376eecf8c759537b6b9aab"} Dec 01 17:55:42 crc kubenswrapper[4868]: I1201 17:55:42.864106 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" Dec 01 17:55:42 crc kubenswrapper[4868]: I1201 17:55:42.916161 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-inventory\") pod \"0d17c1fe-ab49-4510-a2cc-ed17ae06abad\" (UID: \"0d17c1fe-ab49-4510-a2cc-ed17ae06abad\") " Dec 01 17:55:42 crc kubenswrapper[4868]: I1201 17:55:42.916272 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-ssh-key\") pod \"0d17c1fe-ab49-4510-a2cc-ed17ae06abad\" (UID: \"0d17c1fe-ab49-4510-a2cc-ed17ae06abad\") " Dec 01 17:55:42 crc kubenswrapper[4868]: I1201 17:55:42.916354 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rvcc\" (UniqueName: \"kubernetes.io/projected/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-kube-api-access-2rvcc\") pod \"0d17c1fe-ab49-4510-a2cc-ed17ae06abad\" (UID: \"0d17c1fe-ab49-4510-a2cc-ed17ae06abad\") " Dec 01 17:55:42 crc kubenswrapper[4868]: I1201 17:55:42.922902 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-kube-api-access-2rvcc" (OuterVolumeSpecName: "kube-api-access-2rvcc") pod "0d17c1fe-ab49-4510-a2cc-ed17ae06abad" (UID: "0d17c1fe-ab49-4510-a2cc-ed17ae06abad"). InnerVolumeSpecName "kube-api-access-2rvcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:55:42 crc kubenswrapper[4868]: I1201 17:55:42.943131 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-inventory" (OuterVolumeSpecName: "inventory") pod "0d17c1fe-ab49-4510-a2cc-ed17ae06abad" (UID: "0d17c1fe-ab49-4510-a2cc-ed17ae06abad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:55:42 crc kubenswrapper[4868]: I1201 17:55:42.944521 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0d17c1fe-ab49-4510-a2cc-ed17ae06abad" (UID: "0d17c1fe-ab49-4510-a2cc-ed17ae06abad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.018532 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.018574 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rvcc\" (UniqueName: \"kubernetes.io/projected/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-kube-api-access-2rvcc\") on node \"crc\" DevicePath \"\"" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.018588 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d17c1fe-ab49-4510-a2cc-ed17ae06abad-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.438744 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" event={"ID":"0d17c1fe-ab49-4510-a2cc-ed17ae06abad","Type":"ContainerDied","Data":"39d7649e3458017257d401590feb0fd4a71a6c58f31dd990ebed233b6a289d77"} Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.439088 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39d7649e3458017257d401590feb0fd4a71a6c58f31dd990ebed233b6a289d77" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.439150 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.530502 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zjnp8"] Dec 01 17:55:43 crc kubenswrapper[4868]: E1201 17:55:43.530935 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d17c1fe-ab49-4510-a2cc-ed17ae06abad" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.532741 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d17c1fe-ab49-4510-a2cc-ed17ae06abad" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.533087 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d17c1fe-ab49-4510-a2cc-ed17ae06abad" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.533991 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.536705 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.537386 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.537514 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.539239 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.542903 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zjnp8"] Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.628611 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3f7f2561-e242-48b1-a363-50bdecab86de-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-zjnp8\" (UID: \"3f7f2561-e242-48b1-a363-50bdecab86de\") " pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.628679 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3f7f2561-e242-48b1-a363-50bdecab86de-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-zjnp8\" (UID: \"3f7f2561-e242-48b1-a363-50bdecab86de\") " pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.628710 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjfnf\" (UniqueName: \"kubernetes.io/projected/3f7f2561-e242-48b1-a363-50bdecab86de-kube-api-access-xjfnf\") pod \"ssh-known-hosts-edpm-deployment-zjnp8\" (UID: \"3f7f2561-e242-48b1-a363-50bdecab86de\") " pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.731266 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3f7f2561-e242-48b1-a363-50bdecab86de-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-zjnp8\" (UID: \"3f7f2561-e242-48b1-a363-50bdecab86de\") " pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.731336 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3f7f2561-e242-48b1-a363-50bdecab86de-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-zjnp8\" (UID: \"3f7f2561-e242-48b1-a363-50bdecab86de\") " pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.731373 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjfnf\" (UniqueName: \"kubernetes.io/projected/3f7f2561-e242-48b1-a363-50bdecab86de-kube-api-access-xjfnf\") pod \"ssh-known-hosts-edpm-deployment-zjnp8\" (UID: \"3f7f2561-e242-48b1-a363-50bdecab86de\") " pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.737322 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3f7f2561-e242-48b1-a363-50bdecab86de-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-zjnp8\" (UID: \"3f7f2561-e242-48b1-a363-50bdecab86de\") " pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.742445 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3f7f2561-e242-48b1-a363-50bdecab86de-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-zjnp8\" (UID: \"3f7f2561-e242-48b1-a363-50bdecab86de\") " pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.748126 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjfnf\" (UniqueName: \"kubernetes.io/projected/3f7f2561-e242-48b1-a363-50bdecab86de-kube-api-access-xjfnf\") pod \"ssh-known-hosts-edpm-deployment-zjnp8\" (UID: \"3f7f2561-e242-48b1-a363-50bdecab86de\") " pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" Dec 01 17:55:43 crc kubenswrapper[4868]: I1201 17:55:43.866194 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" Dec 01 17:55:44 crc kubenswrapper[4868]: I1201 17:55:44.385560 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zjnp8"] Dec 01 17:55:44 crc kubenswrapper[4868]: I1201 17:55:44.450371 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" event={"ID":"3f7f2561-e242-48b1-a363-50bdecab86de","Type":"ContainerStarted","Data":"2f3c0e23fe1072155135c8d455e729059f107a96679de7c549a2929be7e9d88f"} Dec 01 17:55:45 crc kubenswrapper[4868]: I1201 17:55:45.469154 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" event={"ID":"3f7f2561-e242-48b1-a363-50bdecab86de","Type":"ContainerStarted","Data":"e882db80832f645cb77e51985f3cd09f264596d424411c5ef0e52cc363f77279"} Dec 01 17:55:45 crc kubenswrapper[4868]: I1201 17:55:45.496907 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" podStartSLOduration=2.026455097 podStartE2EDuration="2.496888905s" podCreationTimestamp="2025-12-01 17:55:43 +0000 UTC" firstStartedPulling="2025-12-01 17:55:44.404650988 +0000 UTC m=+1816.775761399" lastFinishedPulling="2025-12-01 17:55:44.875084796 +0000 UTC m=+1817.246195207" observedRunningTime="2025-12-01 17:55:45.488889658 +0000 UTC m=+1817.860000069" watchObservedRunningTime="2025-12-01 17:55:45.496888905 +0000 UTC m=+1817.867999316" Dec 01 17:55:48 crc kubenswrapper[4868]: I1201 17:55:48.178876 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:55:48 crc kubenswrapper[4868]: E1201 17:55:48.181051 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:55:52 crc kubenswrapper[4868]: I1201 17:55:52.537726 4868 generic.go:334] "Generic (PLEG): container finished" podID="3f7f2561-e242-48b1-a363-50bdecab86de" containerID="e882db80832f645cb77e51985f3cd09f264596d424411c5ef0e52cc363f77279" exitCode=0 Dec 01 17:55:52 crc kubenswrapper[4868]: I1201 17:55:52.537815 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" event={"ID":"3f7f2561-e242-48b1-a363-50bdecab86de","Type":"ContainerDied","Data":"e882db80832f645cb77e51985f3cd09f264596d424411c5ef0e52cc363f77279"} Dec 01 17:55:53 crc kubenswrapper[4868]: I1201 17:55:53.952305 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.059790 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjfnf\" (UniqueName: \"kubernetes.io/projected/3f7f2561-e242-48b1-a363-50bdecab86de-kube-api-access-xjfnf\") pod \"3f7f2561-e242-48b1-a363-50bdecab86de\" (UID: \"3f7f2561-e242-48b1-a363-50bdecab86de\") " Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.060265 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3f7f2561-e242-48b1-a363-50bdecab86de-inventory-0\") pod \"3f7f2561-e242-48b1-a363-50bdecab86de\" (UID: \"3f7f2561-e242-48b1-a363-50bdecab86de\") " Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.060730 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3f7f2561-e242-48b1-a363-50bdecab86de-ssh-key-openstack-edpm-ipam\") pod \"3f7f2561-e242-48b1-a363-50bdecab86de\" (UID: \"3f7f2561-e242-48b1-a363-50bdecab86de\") " Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.067245 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f7f2561-e242-48b1-a363-50bdecab86de-kube-api-access-xjfnf" (OuterVolumeSpecName: "kube-api-access-xjfnf") pod "3f7f2561-e242-48b1-a363-50bdecab86de" (UID: "3f7f2561-e242-48b1-a363-50bdecab86de"). InnerVolumeSpecName "kube-api-access-xjfnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.096085 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f7f2561-e242-48b1-a363-50bdecab86de-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3f7f2561-e242-48b1-a363-50bdecab86de" (UID: "3f7f2561-e242-48b1-a363-50bdecab86de"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.096729 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f7f2561-e242-48b1-a363-50bdecab86de-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "3f7f2561-e242-48b1-a363-50bdecab86de" (UID: "3f7f2561-e242-48b1-a363-50bdecab86de"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.163481 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3f7f2561-e242-48b1-a363-50bdecab86de-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.163530 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjfnf\" (UniqueName: \"kubernetes.io/projected/3f7f2561-e242-48b1-a363-50bdecab86de-kube-api-access-xjfnf\") on node \"crc\" DevicePath \"\"" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.163547 4868 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3f7f2561-e242-48b1-a363-50bdecab86de-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.562334 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" event={"ID":"3f7f2561-e242-48b1-a363-50bdecab86de","Type":"ContainerDied","Data":"2f3c0e23fe1072155135c8d455e729059f107a96679de7c549a2929be7e9d88f"} Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.562378 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f3c0e23fe1072155135c8d455e729059f107a96679de7c549a2929be7e9d88f" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.562876 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zjnp8" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.662475 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p"] Dec 01 17:55:54 crc kubenswrapper[4868]: E1201 17:55:54.662924 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f7f2561-e242-48b1-a363-50bdecab86de" containerName="ssh-known-hosts-edpm-deployment" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.662962 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f7f2561-e242-48b1-a363-50bdecab86de" containerName="ssh-known-hosts-edpm-deployment" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.663141 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f7f2561-e242-48b1-a363-50bdecab86de" containerName="ssh-known-hosts-edpm-deployment" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.663913 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.666901 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.667161 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.668302 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.668571 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.674629 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p"] Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.780398 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxd5f\" (UniqueName: \"kubernetes.io/projected/2163d223-acf7-452f-8a2e-5461307cf82e-kube-api-access-nxd5f\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q9z9p\" (UID: \"2163d223-acf7-452f-8a2e-5461307cf82e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.780551 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2163d223-acf7-452f-8a2e-5461307cf82e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q9z9p\" (UID: \"2163d223-acf7-452f-8a2e-5461307cf82e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.780589 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2163d223-acf7-452f-8a2e-5461307cf82e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q9z9p\" (UID: \"2163d223-acf7-452f-8a2e-5461307cf82e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.881978 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2163d223-acf7-452f-8a2e-5461307cf82e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q9z9p\" (UID: \"2163d223-acf7-452f-8a2e-5461307cf82e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.882104 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxd5f\" (UniqueName: \"kubernetes.io/projected/2163d223-acf7-452f-8a2e-5461307cf82e-kube-api-access-nxd5f\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q9z9p\" (UID: \"2163d223-acf7-452f-8a2e-5461307cf82e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.882195 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2163d223-acf7-452f-8a2e-5461307cf82e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q9z9p\" (UID: \"2163d223-acf7-452f-8a2e-5461307cf82e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.885761 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2163d223-acf7-452f-8a2e-5461307cf82e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q9z9p\" (UID: \"2163d223-acf7-452f-8a2e-5461307cf82e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.894662 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2163d223-acf7-452f-8a2e-5461307cf82e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q9z9p\" (UID: \"2163d223-acf7-452f-8a2e-5461307cf82e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.900854 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxd5f\" (UniqueName: \"kubernetes.io/projected/2163d223-acf7-452f-8a2e-5461307cf82e-kube-api-access-nxd5f\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q9z9p\" (UID: \"2163d223-acf7-452f-8a2e-5461307cf82e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" Dec 01 17:55:54 crc kubenswrapper[4868]: I1201 17:55:54.986466 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" Dec 01 17:55:55 crc kubenswrapper[4868]: I1201 17:55:55.516529 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p"] Dec 01 17:55:55 crc kubenswrapper[4868]: I1201 17:55:55.576180 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" event={"ID":"2163d223-acf7-452f-8a2e-5461307cf82e","Type":"ContainerStarted","Data":"7343550acd7e81bbe11b0d9bfdb9dddbfe13dc182d4dc61f47d950e17477592a"} Dec 01 17:55:56 crc kubenswrapper[4868]: I1201 17:55:56.585137 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" event={"ID":"2163d223-acf7-452f-8a2e-5461307cf82e","Type":"ContainerStarted","Data":"7555c1c8e9c909e13ea47725f1e60ad8c564a294bb89bcb667df39d4ea20a52f"} Dec 01 17:55:56 crc kubenswrapper[4868]: I1201 17:55:56.612193 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" podStartSLOduration=1.911922245 podStartE2EDuration="2.61217009s" podCreationTimestamp="2025-12-01 17:55:54 +0000 UTC" firstStartedPulling="2025-12-01 17:55:55.514931318 +0000 UTC m=+1827.886041729" lastFinishedPulling="2025-12-01 17:55:56.215179163 +0000 UTC m=+1828.586289574" observedRunningTime="2025-12-01 17:55:56.599822816 +0000 UTC m=+1828.970933237" watchObservedRunningTime="2025-12-01 17:55:56.61217009 +0000 UTC m=+1828.983280511" Dec 01 17:56:03 crc kubenswrapper[4868]: I1201 17:56:03.173113 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:56:03 crc kubenswrapper[4868]: E1201 17:56:03.174717 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:56:04 crc kubenswrapper[4868]: I1201 17:56:04.664329 4868 generic.go:334] "Generic (PLEG): container finished" podID="2163d223-acf7-452f-8a2e-5461307cf82e" containerID="7555c1c8e9c909e13ea47725f1e60ad8c564a294bb89bcb667df39d4ea20a52f" exitCode=0 Dec 01 17:56:04 crc kubenswrapper[4868]: I1201 17:56:04.664415 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" event={"ID":"2163d223-acf7-452f-8a2e-5461307cf82e","Type":"ContainerDied","Data":"7555c1c8e9c909e13ea47725f1e60ad8c564a294bb89bcb667df39d4ea20a52f"} Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.096896 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.256314 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxd5f\" (UniqueName: \"kubernetes.io/projected/2163d223-acf7-452f-8a2e-5461307cf82e-kube-api-access-nxd5f\") pod \"2163d223-acf7-452f-8a2e-5461307cf82e\" (UID: \"2163d223-acf7-452f-8a2e-5461307cf82e\") " Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.256476 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2163d223-acf7-452f-8a2e-5461307cf82e-ssh-key\") pod \"2163d223-acf7-452f-8a2e-5461307cf82e\" (UID: \"2163d223-acf7-452f-8a2e-5461307cf82e\") " Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.256536 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2163d223-acf7-452f-8a2e-5461307cf82e-inventory\") pod \"2163d223-acf7-452f-8a2e-5461307cf82e\" (UID: \"2163d223-acf7-452f-8a2e-5461307cf82e\") " Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.263216 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2163d223-acf7-452f-8a2e-5461307cf82e-kube-api-access-nxd5f" (OuterVolumeSpecName: "kube-api-access-nxd5f") pod "2163d223-acf7-452f-8a2e-5461307cf82e" (UID: "2163d223-acf7-452f-8a2e-5461307cf82e"). InnerVolumeSpecName "kube-api-access-nxd5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.287457 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2163d223-acf7-452f-8a2e-5461307cf82e-inventory" (OuterVolumeSpecName: "inventory") pod "2163d223-acf7-452f-8a2e-5461307cf82e" (UID: "2163d223-acf7-452f-8a2e-5461307cf82e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.293706 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2163d223-acf7-452f-8a2e-5461307cf82e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2163d223-acf7-452f-8a2e-5461307cf82e" (UID: "2163d223-acf7-452f-8a2e-5461307cf82e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.359091 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxd5f\" (UniqueName: \"kubernetes.io/projected/2163d223-acf7-452f-8a2e-5461307cf82e-kube-api-access-nxd5f\") on node \"crc\" DevicePath \"\"" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.359135 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2163d223-acf7-452f-8a2e-5461307cf82e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.359150 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2163d223-acf7-452f-8a2e-5461307cf82e-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.686154 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" event={"ID":"2163d223-acf7-452f-8a2e-5461307cf82e","Type":"ContainerDied","Data":"7343550acd7e81bbe11b0d9bfdb9dddbfe13dc182d4dc61f47d950e17477592a"} Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.686204 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7343550acd7e81bbe11b0d9bfdb9dddbfe13dc182d4dc61f47d950e17477592a" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.686258 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q9z9p" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.822432 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9"] Dec 01 17:56:06 crc kubenswrapper[4868]: E1201 17:56:06.822832 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2163d223-acf7-452f-8a2e-5461307cf82e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.822849 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="2163d223-acf7-452f-8a2e-5461307cf82e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.823060 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="2163d223-acf7-452f-8a2e-5461307cf82e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.823761 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.825970 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.829125 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.829298 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.829533 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.836853 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9"] Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.968653 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17fd50a0-00e4-4b51-9631-30b48bd3013f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9\" (UID: \"17fd50a0-00e4-4b51-9631-30b48bd3013f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.968700 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwhqs\" (UniqueName: \"kubernetes.io/projected/17fd50a0-00e4-4b51-9631-30b48bd3013f-kube-api-access-nwhqs\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9\" (UID: \"17fd50a0-00e4-4b51-9631-30b48bd3013f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" Dec 01 17:56:06 crc kubenswrapper[4868]: I1201 17:56:06.968725 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17fd50a0-00e4-4b51-9631-30b48bd3013f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9\" (UID: \"17fd50a0-00e4-4b51-9631-30b48bd3013f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" Dec 01 17:56:07 crc kubenswrapper[4868]: I1201 17:56:07.070471 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17fd50a0-00e4-4b51-9631-30b48bd3013f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9\" (UID: \"17fd50a0-00e4-4b51-9631-30b48bd3013f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" Dec 01 17:56:07 crc kubenswrapper[4868]: I1201 17:56:07.070530 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwhqs\" (UniqueName: \"kubernetes.io/projected/17fd50a0-00e4-4b51-9631-30b48bd3013f-kube-api-access-nwhqs\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9\" (UID: \"17fd50a0-00e4-4b51-9631-30b48bd3013f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" Dec 01 17:56:07 crc kubenswrapper[4868]: I1201 17:56:07.070552 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17fd50a0-00e4-4b51-9631-30b48bd3013f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9\" (UID: \"17fd50a0-00e4-4b51-9631-30b48bd3013f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" Dec 01 17:56:07 crc kubenswrapper[4868]: I1201 17:56:07.075978 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17fd50a0-00e4-4b51-9631-30b48bd3013f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9\" (UID: \"17fd50a0-00e4-4b51-9631-30b48bd3013f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" Dec 01 17:56:07 crc kubenswrapper[4868]: I1201 17:56:07.076143 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17fd50a0-00e4-4b51-9631-30b48bd3013f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9\" (UID: \"17fd50a0-00e4-4b51-9631-30b48bd3013f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" Dec 01 17:56:07 crc kubenswrapper[4868]: I1201 17:56:07.088118 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwhqs\" (UniqueName: \"kubernetes.io/projected/17fd50a0-00e4-4b51-9631-30b48bd3013f-kube-api-access-nwhqs\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9\" (UID: \"17fd50a0-00e4-4b51-9631-30b48bd3013f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" Dec 01 17:56:07 crc kubenswrapper[4868]: I1201 17:56:07.141006 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" Dec 01 17:56:07 crc kubenswrapper[4868]: I1201 17:56:07.715095 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9"] Dec 01 17:56:08 crc kubenswrapper[4868]: I1201 17:56:08.705161 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" event={"ID":"17fd50a0-00e4-4b51-9631-30b48bd3013f","Type":"ContainerStarted","Data":"cc194a1551a43135d18a32cb3f3ba782ebdc4a6b9d7f7da779f18c3ca353ad62"} Dec 01 17:56:09 crc kubenswrapper[4868]: I1201 17:56:09.716116 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" event={"ID":"17fd50a0-00e4-4b51-9631-30b48bd3013f","Type":"ContainerStarted","Data":"d497e099c592ed5f37f58dc0d17f47b0f7b9e3a17fabae1896743d4e83b307fd"} Dec 01 17:56:09 crc kubenswrapper[4868]: I1201 17:56:09.749634 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" podStartSLOduration=2.625355979 podStartE2EDuration="3.749587384s" podCreationTimestamp="2025-12-01 17:56:06 +0000 UTC" firstStartedPulling="2025-12-01 17:56:07.719684827 +0000 UTC m=+1840.090795258" lastFinishedPulling="2025-12-01 17:56:08.843916212 +0000 UTC m=+1841.215026663" observedRunningTime="2025-12-01 17:56:09.745702589 +0000 UTC m=+1842.116813010" watchObservedRunningTime="2025-12-01 17:56:09.749587384 +0000 UTC m=+1842.120697795" Dec 01 17:56:18 crc kubenswrapper[4868]: I1201 17:56:18.187827 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:56:18 crc kubenswrapper[4868]: E1201 17:56:18.189978 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:56:18 crc kubenswrapper[4868]: I1201 17:56:18.822502 4868 generic.go:334] "Generic (PLEG): container finished" podID="17fd50a0-00e4-4b51-9631-30b48bd3013f" containerID="d497e099c592ed5f37f58dc0d17f47b0f7b9e3a17fabae1896743d4e83b307fd" exitCode=0 Dec 01 17:56:18 crc kubenswrapper[4868]: I1201 17:56:18.822583 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" event={"ID":"17fd50a0-00e4-4b51-9631-30b48bd3013f","Type":"ContainerDied","Data":"d497e099c592ed5f37f58dc0d17f47b0f7b9e3a17fabae1896743d4e83b307fd"} Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.063806 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-pjj7m"] Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.075738 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-pjj7m"] Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.197971 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="491393ef-ffbb-4a81-a477-5856765b7a48" path="/var/lib/kubelet/pods/491393ef-ffbb-4a81-a477-5856765b7a48/volumes" Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.290542 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.337909 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwhqs\" (UniqueName: \"kubernetes.io/projected/17fd50a0-00e4-4b51-9631-30b48bd3013f-kube-api-access-nwhqs\") pod \"17fd50a0-00e4-4b51-9631-30b48bd3013f\" (UID: \"17fd50a0-00e4-4b51-9631-30b48bd3013f\") " Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.338629 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17fd50a0-00e4-4b51-9631-30b48bd3013f-ssh-key\") pod \"17fd50a0-00e4-4b51-9631-30b48bd3013f\" (UID: \"17fd50a0-00e4-4b51-9631-30b48bd3013f\") " Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.339128 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17fd50a0-00e4-4b51-9631-30b48bd3013f-inventory\") pod \"17fd50a0-00e4-4b51-9631-30b48bd3013f\" (UID: \"17fd50a0-00e4-4b51-9631-30b48bd3013f\") " Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.346231 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17fd50a0-00e4-4b51-9631-30b48bd3013f-kube-api-access-nwhqs" (OuterVolumeSpecName: "kube-api-access-nwhqs") pod "17fd50a0-00e4-4b51-9631-30b48bd3013f" (UID: "17fd50a0-00e4-4b51-9631-30b48bd3013f"). InnerVolumeSpecName "kube-api-access-nwhqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.375331 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17fd50a0-00e4-4b51-9631-30b48bd3013f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "17fd50a0-00e4-4b51-9631-30b48bd3013f" (UID: "17fd50a0-00e4-4b51-9631-30b48bd3013f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.379125 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17fd50a0-00e4-4b51-9631-30b48bd3013f-inventory" (OuterVolumeSpecName: "inventory") pod "17fd50a0-00e4-4b51-9631-30b48bd3013f" (UID: "17fd50a0-00e4-4b51-9631-30b48bd3013f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.445721 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwhqs\" (UniqueName: \"kubernetes.io/projected/17fd50a0-00e4-4b51-9631-30b48bd3013f-kube-api-access-nwhqs\") on node \"crc\" DevicePath \"\"" Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.445762 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17fd50a0-00e4-4b51-9631-30b48bd3013f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.445774 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17fd50a0-00e4-4b51-9631-30b48bd3013f-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.842316 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" event={"ID":"17fd50a0-00e4-4b51-9631-30b48bd3013f","Type":"ContainerDied","Data":"cc194a1551a43135d18a32cb3f3ba782ebdc4a6b9d7f7da779f18c3ca353ad62"} Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.842801 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc194a1551a43135d18a32cb3f3ba782ebdc4a6b9d7f7da779f18c3ca353ad62" Dec 01 17:56:20 crc kubenswrapper[4868]: I1201 17:56:20.842389 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.041576 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph"] Dec 01 17:56:21 crc kubenswrapper[4868]: E1201 17:56:21.042061 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17fd50a0-00e4-4b51-9631-30b48bd3013f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.042082 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="17fd50a0-00e4-4b51-9631-30b48bd3013f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.042295 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="17fd50a0-00e4-4b51-9631-30b48bd3013f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.043006 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.046023 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.046149 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.046180 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.046248 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.046363 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.046524 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.046531 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.047283 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.062662 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph"] Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.158316 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.158383 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.158414 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.158436 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.158472 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.158503 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.158540 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.158569 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.158598 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.158615 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.158653 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.158676 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.158696 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.158717 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgc95\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-kube-api-access-zgc95\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.260047 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.260144 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.260178 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.260204 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.260378 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.261062 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.261109 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.261165 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.261205 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.261229 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.261260 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.261285 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.261313 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.261338 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgc95\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-kube-api-access-zgc95\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.268151 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.268509 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.268590 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.268697 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.268687 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.269466 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.270673 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.270889 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.272427 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.272655 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.273209 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.280065 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.280571 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.286902 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgc95\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-kube-api-access-zgc95\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-drbph\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.397459 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:56:21 crc kubenswrapper[4868]: I1201 17:56:21.993013 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph"] Dec 01 17:56:22 crc kubenswrapper[4868]: I1201 17:56:22.009112 4868 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 17:56:22 crc kubenswrapper[4868]: I1201 17:56:22.866851 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" event={"ID":"31db6887-0dfe-44c1-8951-757f2287cd9e","Type":"ContainerStarted","Data":"d488bb0f164a17913ffa46fb60470bbe59136a610df561ab36a880971a51ebd4"} Dec 01 17:56:22 crc kubenswrapper[4868]: I1201 17:56:22.867719 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" event={"ID":"31db6887-0dfe-44c1-8951-757f2287cd9e","Type":"ContainerStarted","Data":"406c92c8983d48c51c7c9bfc83c513a961c7730b2c3e4cc272be4d218cca819d"} Dec 01 17:56:22 crc kubenswrapper[4868]: I1201 17:56:22.899153 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" podStartSLOduration=1.471971978 podStartE2EDuration="1.899121694s" podCreationTimestamp="2025-12-01 17:56:21 +0000 UTC" firstStartedPulling="2025-12-01 17:56:22.008838379 +0000 UTC m=+1854.379948790" lastFinishedPulling="2025-12-01 17:56:22.435988075 +0000 UTC m=+1854.807098506" observedRunningTime="2025-12-01 17:56:22.885767012 +0000 UTC m=+1855.256877423" watchObservedRunningTime="2025-12-01 17:56:22.899121694 +0000 UTC m=+1855.270232105" Dec 01 17:56:31 crc kubenswrapper[4868]: I1201 17:56:31.172576 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:56:31 crc kubenswrapper[4868]: E1201 17:56:31.173673 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:56:41 crc kubenswrapper[4868]: I1201 17:56:41.141085 4868 scope.go:117] "RemoveContainer" containerID="98997845d571b01016aa693a291d1d21d2117a71f509381ce801cc152937438a" Dec 01 17:56:44 crc kubenswrapper[4868]: I1201 17:56:44.172501 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:56:44 crc kubenswrapper[4868]: E1201 17:56:44.173578 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:56:55 crc kubenswrapper[4868]: I1201 17:56:55.172778 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:56:55 crc kubenswrapper[4868]: E1201 17:56:55.173671 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:56:59 crc kubenswrapper[4868]: I1201 17:56:59.208206 4868 generic.go:334] "Generic (PLEG): container finished" podID="31db6887-0dfe-44c1-8951-757f2287cd9e" containerID="d488bb0f164a17913ffa46fb60470bbe59136a610df561ab36a880971a51ebd4" exitCode=0 Dec 01 17:56:59 crc kubenswrapper[4868]: I1201 17:56:59.208300 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" event={"ID":"31db6887-0dfe-44c1-8951-757f2287cd9e","Type":"ContainerDied","Data":"d488bb0f164a17913ffa46fb60470bbe59136a610df561ab36a880971a51ebd4"} Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.611870 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.749763 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-neutron-metadata-combined-ca-bundle\") pod \"31db6887-0dfe-44c1-8951-757f2287cd9e\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.749833 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-bootstrap-combined-ca-bundle\") pod \"31db6887-0dfe-44c1-8951-757f2287cd9e\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.749863 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-telemetry-combined-ca-bundle\") pod \"31db6887-0dfe-44c1-8951-757f2287cd9e\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.749903 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-repo-setup-combined-ca-bundle\") pod \"31db6887-0dfe-44c1-8951-757f2287cd9e\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.749932 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-inventory\") pod \"31db6887-0dfe-44c1-8951-757f2287cd9e\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.749974 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"31db6887-0dfe-44c1-8951-757f2287cd9e\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.750003 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"31db6887-0dfe-44c1-8951-757f2287cd9e\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.750023 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"31db6887-0dfe-44c1-8951-757f2287cd9e\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.750063 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-ssh-key\") pod \"31db6887-0dfe-44c1-8951-757f2287cd9e\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.750126 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-ovn-combined-ca-bundle\") pod \"31db6887-0dfe-44c1-8951-757f2287cd9e\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.750149 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-libvirt-combined-ca-bundle\") pod \"31db6887-0dfe-44c1-8951-757f2287cd9e\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.750183 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-nova-combined-ca-bundle\") pod \"31db6887-0dfe-44c1-8951-757f2287cd9e\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.750272 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"31db6887-0dfe-44c1-8951-757f2287cd9e\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.750293 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgc95\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-kube-api-access-zgc95\") pod \"31db6887-0dfe-44c1-8951-757f2287cd9e\" (UID: \"31db6887-0dfe-44c1-8951-757f2287cd9e\") " Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.757491 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "31db6887-0dfe-44c1-8951-757f2287cd9e" (UID: "31db6887-0dfe-44c1-8951-757f2287cd9e"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.758462 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "31db6887-0dfe-44c1-8951-757f2287cd9e" (UID: "31db6887-0dfe-44c1-8951-757f2287cd9e"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.758527 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "31db6887-0dfe-44c1-8951-757f2287cd9e" (UID: "31db6887-0dfe-44c1-8951-757f2287cd9e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.758565 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "31db6887-0dfe-44c1-8951-757f2287cd9e" (UID: "31db6887-0dfe-44c1-8951-757f2287cd9e"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.758595 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "31db6887-0dfe-44c1-8951-757f2287cd9e" (UID: "31db6887-0dfe-44c1-8951-757f2287cd9e"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.759184 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "31db6887-0dfe-44c1-8951-757f2287cd9e" (UID: "31db6887-0dfe-44c1-8951-757f2287cd9e"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.759496 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-kube-api-access-zgc95" (OuterVolumeSpecName: "kube-api-access-zgc95") pod "31db6887-0dfe-44c1-8951-757f2287cd9e" (UID: "31db6887-0dfe-44c1-8951-757f2287cd9e"). InnerVolumeSpecName "kube-api-access-zgc95". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.759725 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "31db6887-0dfe-44c1-8951-757f2287cd9e" (UID: "31db6887-0dfe-44c1-8951-757f2287cd9e"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.760145 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "31db6887-0dfe-44c1-8951-757f2287cd9e" (UID: "31db6887-0dfe-44c1-8951-757f2287cd9e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.760673 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "31db6887-0dfe-44c1-8951-757f2287cd9e" (UID: "31db6887-0dfe-44c1-8951-757f2287cd9e"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.762483 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "31db6887-0dfe-44c1-8951-757f2287cd9e" (UID: "31db6887-0dfe-44c1-8951-757f2287cd9e"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.762729 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "31db6887-0dfe-44c1-8951-757f2287cd9e" (UID: "31db6887-0dfe-44c1-8951-757f2287cd9e"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.783163 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "31db6887-0dfe-44c1-8951-757f2287cd9e" (UID: "31db6887-0dfe-44c1-8951-757f2287cd9e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.785288 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-inventory" (OuterVolumeSpecName: "inventory") pod "31db6887-0dfe-44c1-8951-757f2287cd9e" (UID: "31db6887-0dfe-44c1-8951-757f2287cd9e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.852370 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.852411 4868 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.852427 4868 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.852440 4868 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.852453 4868 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.852465 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgc95\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-kube-api-access-zgc95\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.852477 4868 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.852490 4868 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.852502 4868 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.852514 4868 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.852525 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31db6887-0dfe-44c1-8951-757f2287cd9e-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.852537 4868 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.852549 4868 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:00 crc kubenswrapper[4868]: I1201 17:57:00.852564 4868 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/31db6887-0dfe-44c1-8951-757f2287cd9e-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.233857 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" event={"ID":"31db6887-0dfe-44c1-8951-757f2287cd9e","Type":"ContainerDied","Data":"406c92c8983d48c51c7c9bfc83c513a961c7730b2c3e4cc272be4d218cca819d"} Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.234216 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="406c92c8983d48c51c7c9bfc83c513a961c7730b2c3e4cc272be4d218cca819d" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.234003 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-drbph" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.334866 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w"] Dec 01 17:57:01 crc kubenswrapper[4868]: E1201 17:57:01.335437 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31db6887-0dfe-44c1-8951-757f2287cd9e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.335463 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="31db6887-0dfe-44c1-8951-757f2287cd9e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.335751 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="31db6887-0dfe-44c1-8951-757f2287cd9e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.336635 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.338814 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.338933 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.339045 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.339142 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.339709 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.346864 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w"] Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.462227 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzwnb\" (UniqueName: \"kubernetes.io/projected/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-kube-api-access-kzwnb\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nmn4w\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.462302 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nmn4w\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.462345 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nmn4w\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.462571 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nmn4w\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.462847 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nmn4w\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.564514 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzwnb\" (UniqueName: \"kubernetes.io/projected/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-kube-api-access-kzwnb\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nmn4w\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.564587 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nmn4w\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.564638 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nmn4w\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.564719 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nmn4w\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.564825 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nmn4w\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.565782 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nmn4w\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.569660 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nmn4w\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.569994 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nmn4w\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.571036 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nmn4w\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.580397 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzwnb\" (UniqueName: \"kubernetes.io/projected/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-kube-api-access-kzwnb\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nmn4w\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:01 crc kubenswrapper[4868]: I1201 17:57:01.659573 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:57:02 crc kubenswrapper[4868]: I1201 17:57:02.187286 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w"] Dec 01 17:57:02 crc kubenswrapper[4868]: I1201 17:57:02.242564 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" event={"ID":"216fe1b0-ec90-4ee7-91e4-aa24476b39b4","Type":"ContainerStarted","Data":"9f697d314c6af4badc76ddde68481627d1178c986adc919a4c56c0ec424e04f3"} Dec 01 17:57:03 crc kubenswrapper[4868]: I1201 17:57:03.254501 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" event={"ID":"216fe1b0-ec90-4ee7-91e4-aa24476b39b4","Type":"ContainerStarted","Data":"6c6841afcf09b5434a25c5bbcd7c972d30ea65862b279f445f00e658538b6538"} Dec 01 17:57:03 crc kubenswrapper[4868]: I1201 17:57:03.286650 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" podStartSLOduration=1.487479652 podStartE2EDuration="2.286623822s" podCreationTimestamp="2025-12-01 17:57:01 +0000 UTC" firstStartedPulling="2025-12-01 17:57:02.189828923 +0000 UTC m=+1894.560939334" lastFinishedPulling="2025-12-01 17:57:02.988973093 +0000 UTC m=+1895.360083504" observedRunningTime="2025-12-01 17:57:03.279123238 +0000 UTC m=+1895.650233649" watchObservedRunningTime="2025-12-01 17:57:03.286623822 +0000 UTC m=+1895.657734233" Dec 01 17:57:10 crc kubenswrapper[4868]: I1201 17:57:10.172958 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:57:10 crc kubenswrapper[4868]: E1201 17:57:10.174797 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:57:20 crc kubenswrapper[4868]: I1201 17:57:20.298405 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s8hl5"] Dec 01 17:57:20 crc kubenswrapper[4868]: I1201 17:57:20.301015 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:20 crc kubenswrapper[4868]: I1201 17:57:20.322838 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s8hl5"] Dec 01 17:57:20 crc kubenswrapper[4868]: I1201 17:57:20.347551 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hzlj\" (UniqueName: \"kubernetes.io/projected/6549a036-6568-40dd-b26d-f1b91357e550-kube-api-access-4hzlj\") pod \"certified-operators-s8hl5\" (UID: \"6549a036-6568-40dd-b26d-f1b91357e550\") " pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:20 crc kubenswrapper[4868]: I1201 17:57:20.347618 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6549a036-6568-40dd-b26d-f1b91357e550-catalog-content\") pod \"certified-operators-s8hl5\" (UID: \"6549a036-6568-40dd-b26d-f1b91357e550\") " pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:20 crc kubenswrapper[4868]: I1201 17:57:20.347924 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6549a036-6568-40dd-b26d-f1b91357e550-utilities\") pod \"certified-operators-s8hl5\" (UID: \"6549a036-6568-40dd-b26d-f1b91357e550\") " pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:20 crc kubenswrapper[4868]: I1201 17:57:20.450480 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6549a036-6568-40dd-b26d-f1b91357e550-catalog-content\") pod \"certified-operators-s8hl5\" (UID: \"6549a036-6568-40dd-b26d-f1b91357e550\") " pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:20 crc kubenswrapper[4868]: I1201 17:57:20.450581 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6549a036-6568-40dd-b26d-f1b91357e550-utilities\") pod \"certified-operators-s8hl5\" (UID: \"6549a036-6568-40dd-b26d-f1b91357e550\") " pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:20 crc kubenswrapper[4868]: I1201 17:57:20.450690 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hzlj\" (UniqueName: \"kubernetes.io/projected/6549a036-6568-40dd-b26d-f1b91357e550-kube-api-access-4hzlj\") pod \"certified-operators-s8hl5\" (UID: \"6549a036-6568-40dd-b26d-f1b91357e550\") " pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:20 crc kubenswrapper[4868]: I1201 17:57:20.451040 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6549a036-6568-40dd-b26d-f1b91357e550-catalog-content\") pod \"certified-operators-s8hl5\" (UID: \"6549a036-6568-40dd-b26d-f1b91357e550\") " pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:20 crc kubenswrapper[4868]: I1201 17:57:20.451155 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6549a036-6568-40dd-b26d-f1b91357e550-utilities\") pod \"certified-operators-s8hl5\" (UID: \"6549a036-6568-40dd-b26d-f1b91357e550\") " pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:20 crc kubenswrapper[4868]: I1201 17:57:20.471339 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hzlj\" (UniqueName: \"kubernetes.io/projected/6549a036-6568-40dd-b26d-f1b91357e550-kube-api-access-4hzlj\") pod \"certified-operators-s8hl5\" (UID: \"6549a036-6568-40dd-b26d-f1b91357e550\") " pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:20 crc kubenswrapper[4868]: I1201 17:57:20.650465 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:21 crc kubenswrapper[4868]: I1201 17:57:21.155902 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s8hl5"] Dec 01 17:57:21 crc kubenswrapper[4868]: I1201 17:57:21.394721 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s8hl5" event={"ID":"6549a036-6568-40dd-b26d-f1b91357e550","Type":"ContainerStarted","Data":"9f9f82cf1dc197085442ee7d5922864c70d55ee53da1afd9957565cbffcc8d3c"} Dec 01 17:57:21 crc kubenswrapper[4868]: E1201 17:57:21.616959 4868 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6549a036_6568_40dd_b26d_f1b91357e550.slice/crio-conmon-b3f08328fa7d5fe74e2f494fb8fab2b8255a69ce6cdaf5dc37f098b91192e832.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6549a036_6568_40dd_b26d_f1b91357e550.slice/crio-b3f08328fa7d5fe74e2f494fb8fab2b8255a69ce6cdaf5dc37f098b91192e832.scope\": RecentStats: unable to find data in memory cache]" Dec 01 17:57:22 crc kubenswrapper[4868]: I1201 17:57:22.405854 4868 generic.go:334] "Generic (PLEG): container finished" podID="6549a036-6568-40dd-b26d-f1b91357e550" containerID="b3f08328fa7d5fe74e2f494fb8fab2b8255a69ce6cdaf5dc37f098b91192e832" exitCode=0 Dec 01 17:57:22 crc kubenswrapper[4868]: I1201 17:57:22.405965 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s8hl5" event={"ID":"6549a036-6568-40dd-b26d-f1b91357e550","Type":"ContainerDied","Data":"b3f08328fa7d5fe74e2f494fb8fab2b8255a69ce6cdaf5dc37f098b91192e832"} Dec 01 17:57:24 crc kubenswrapper[4868]: I1201 17:57:24.424878 4868 generic.go:334] "Generic (PLEG): container finished" podID="6549a036-6568-40dd-b26d-f1b91357e550" containerID="17cf1266cc598849a8cff9defd067785e895f6d1be32689d5a4d557b3bb82261" exitCode=0 Dec 01 17:57:24 crc kubenswrapper[4868]: I1201 17:57:24.425012 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s8hl5" event={"ID":"6549a036-6568-40dd-b26d-f1b91357e550","Type":"ContainerDied","Data":"17cf1266cc598849a8cff9defd067785e895f6d1be32689d5a4d557b3bb82261"} Dec 01 17:57:25 crc kubenswrapper[4868]: I1201 17:57:25.172675 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:57:25 crc kubenswrapper[4868]: E1201 17:57:25.173324 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:57:25 crc kubenswrapper[4868]: I1201 17:57:25.446610 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s8hl5" event={"ID":"6549a036-6568-40dd-b26d-f1b91357e550","Type":"ContainerStarted","Data":"4edb1aee11b1814072c6ff2d80d1a3f2f52912ff4606588e299e8ccf21886503"} Dec 01 17:57:25 crc kubenswrapper[4868]: I1201 17:57:25.484004 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s8hl5" podStartSLOduration=2.978788245 podStartE2EDuration="5.483979138s" podCreationTimestamp="2025-12-01 17:57:20 +0000 UTC" firstStartedPulling="2025-12-01 17:57:22.408507607 +0000 UTC m=+1914.779618018" lastFinishedPulling="2025-12-01 17:57:24.9136985 +0000 UTC m=+1917.284808911" observedRunningTime="2025-12-01 17:57:25.477895342 +0000 UTC m=+1917.849005763" watchObservedRunningTime="2025-12-01 17:57:25.483979138 +0000 UTC m=+1917.855089549" Dec 01 17:57:30 crc kubenswrapper[4868]: I1201 17:57:30.651089 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:30 crc kubenswrapper[4868]: I1201 17:57:30.651668 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:30 crc kubenswrapper[4868]: I1201 17:57:30.698738 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:31 crc kubenswrapper[4868]: I1201 17:57:31.547129 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:31 crc kubenswrapper[4868]: I1201 17:57:31.596986 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s8hl5"] Dec 01 17:57:33 crc kubenswrapper[4868]: I1201 17:57:33.516226 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s8hl5" podUID="6549a036-6568-40dd-b26d-f1b91357e550" containerName="registry-server" containerID="cri-o://4edb1aee11b1814072c6ff2d80d1a3f2f52912ff4606588e299e8ccf21886503" gracePeriod=2 Dec 01 17:57:33 crc kubenswrapper[4868]: I1201 17:57:33.945089 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.045398 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hzlj\" (UniqueName: \"kubernetes.io/projected/6549a036-6568-40dd-b26d-f1b91357e550-kube-api-access-4hzlj\") pod \"6549a036-6568-40dd-b26d-f1b91357e550\" (UID: \"6549a036-6568-40dd-b26d-f1b91357e550\") " Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.045496 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6549a036-6568-40dd-b26d-f1b91357e550-catalog-content\") pod \"6549a036-6568-40dd-b26d-f1b91357e550\" (UID: \"6549a036-6568-40dd-b26d-f1b91357e550\") " Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.045542 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6549a036-6568-40dd-b26d-f1b91357e550-utilities\") pod \"6549a036-6568-40dd-b26d-f1b91357e550\" (UID: \"6549a036-6568-40dd-b26d-f1b91357e550\") " Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.046538 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6549a036-6568-40dd-b26d-f1b91357e550-utilities" (OuterVolumeSpecName: "utilities") pod "6549a036-6568-40dd-b26d-f1b91357e550" (UID: "6549a036-6568-40dd-b26d-f1b91357e550"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.056208 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6549a036-6568-40dd-b26d-f1b91357e550-kube-api-access-4hzlj" (OuterVolumeSpecName: "kube-api-access-4hzlj") pod "6549a036-6568-40dd-b26d-f1b91357e550" (UID: "6549a036-6568-40dd-b26d-f1b91357e550"). InnerVolumeSpecName "kube-api-access-4hzlj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.092277 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6549a036-6568-40dd-b26d-f1b91357e550-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6549a036-6568-40dd-b26d-f1b91357e550" (UID: "6549a036-6568-40dd-b26d-f1b91357e550"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.148057 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6549a036-6568-40dd-b26d-f1b91357e550-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.148089 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6549a036-6568-40dd-b26d-f1b91357e550-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.148101 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hzlj\" (UniqueName: \"kubernetes.io/projected/6549a036-6568-40dd-b26d-f1b91357e550-kube-api-access-4hzlj\") on node \"crc\" DevicePath \"\"" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.527860 4868 generic.go:334] "Generic (PLEG): container finished" podID="6549a036-6568-40dd-b26d-f1b91357e550" containerID="4edb1aee11b1814072c6ff2d80d1a3f2f52912ff4606588e299e8ccf21886503" exitCode=0 Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.527979 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s8hl5" event={"ID":"6549a036-6568-40dd-b26d-f1b91357e550","Type":"ContainerDied","Data":"4edb1aee11b1814072c6ff2d80d1a3f2f52912ff4606588e299e8ccf21886503"} Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.528132 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s8hl5" event={"ID":"6549a036-6568-40dd-b26d-f1b91357e550","Type":"ContainerDied","Data":"9f9f82cf1dc197085442ee7d5922864c70d55ee53da1afd9957565cbffcc8d3c"} Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.528153 4868 scope.go:117] "RemoveContainer" containerID="4edb1aee11b1814072c6ff2d80d1a3f2f52912ff4606588e299e8ccf21886503" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.527982 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s8hl5" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.556215 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s8hl5"] Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.561041 4868 scope.go:117] "RemoveContainer" containerID="17cf1266cc598849a8cff9defd067785e895f6d1be32689d5a4d557b3bb82261" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.568858 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s8hl5"] Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.581893 4868 scope.go:117] "RemoveContainer" containerID="b3f08328fa7d5fe74e2f494fb8fab2b8255a69ce6cdaf5dc37f098b91192e832" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.622235 4868 scope.go:117] "RemoveContainer" containerID="4edb1aee11b1814072c6ff2d80d1a3f2f52912ff4606588e299e8ccf21886503" Dec 01 17:57:34 crc kubenswrapper[4868]: E1201 17:57:34.622705 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4edb1aee11b1814072c6ff2d80d1a3f2f52912ff4606588e299e8ccf21886503\": container with ID starting with 4edb1aee11b1814072c6ff2d80d1a3f2f52912ff4606588e299e8ccf21886503 not found: ID does not exist" containerID="4edb1aee11b1814072c6ff2d80d1a3f2f52912ff4606588e299e8ccf21886503" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.622745 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4edb1aee11b1814072c6ff2d80d1a3f2f52912ff4606588e299e8ccf21886503"} err="failed to get container status \"4edb1aee11b1814072c6ff2d80d1a3f2f52912ff4606588e299e8ccf21886503\": rpc error: code = NotFound desc = could not find container \"4edb1aee11b1814072c6ff2d80d1a3f2f52912ff4606588e299e8ccf21886503\": container with ID starting with 4edb1aee11b1814072c6ff2d80d1a3f2f52912ff4606588e299e8ccf21886503 not found: ID does not exist" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.622775 4868 scope.go:117] "RemoveContainer" containerID="17cf1266cc598849a8cff9defd067785e895f6d1be32689d5a4d557b3bb82261" Dec 01 17:57:34 crc kubenswrapper[4868]: E1201 17:57:34.623208 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17cf1266cc598849a8cff9defd067785e895f6d1be32689d5a4d557b3bb82261\": container with ID starting with 17cf1266cc598849a8cff9defd067785e895f6d1be32689d5a4d557b3bb82261 not found: ID does not exist" containerID="17cf1266cc598849a8cff9defd067785e895f6d1be32689d5a4d557b3bb82261" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.623232 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17cf1266cc598849a8cff9defd067785e895f6d1be32689d5a4d557b3bb82261"} err="failed to get container status \"17cf1266cc598849a8cff9defd067785e895f6d1be32689d5a4d557b3bb82261\": rpc error: code = NotFound desc = could not find container \"17cf1266cc598849a8cff9defd067785e895f6d1be32689d5a4d557b3bb82261\": container with ID starting with 17cf1266cc598849a8cff9defd067785e895f6d1be32689d5a4d557b3bb82261 not found: ID does not exist" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.623250 4868 scope.go:117] "RemoveContainer" containerID="b3f08328fa7d5fe74e2f494fb8fab2b8255a69ce6cdaf5dc37f098b91192e832" Dec 01 17:57:34 crc kubenswrapper[4868]: E1201 17:57:34.623475 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3f08328fa7d5fe74e2f494fb8fab2b8255a69ce6cdaf5dc37f098b91192e832\": container with ID starting with b3f08328fa7d5fe74e2f494fb8fab2b8255a69ce6cdaf5dc37f098b91192e832 not found: ID does not exist" containerID="b3f08328fa7d5fe74e2f494fb8fab2b8255a69ce6cdaf5dc37f098b91192e832" Dec 01 17:57:34 crc kubenswrapper[4868]: I1201 17:57:34.623499 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3f08328fa7d5fe74e2f494fb8fab2b8255a69ce6cdaf5dc37f098b91192e832"} err="failed to get container status \"b3f08328fa7d5fe74e2f494fb8fab2b8255a69ce6cdaf5dc37f098b91192e832\": rpc error: code = NotFound desc = could not find container \"b3f08328fa7d5fe74e2f494fb8fab2b8255a69ce6cdaf5dc37f098b91192e832\": container with ID starting with b3f08328fa7d5fe74e2f494fb8fab2b8255a69ce6cdaf5dc37f098b91192e832 not found: ID does not exist" Dec 01 17:57:36 crc kubenswrapper[4868]: I1201 17:57:36.172252 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:57:36 crc kubenswrapper[4868]: E1201 17:57:36.173499 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:57:36 crc kubenswrapper[4868]: I1201 17:57:36.185424 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6549a036-6568-40dd-b26d-f1b91357e550" path="/var/lib/kubelet/pods/6549a036-6568-40dd-b26d-f1b91357e550/volumes" Dec 01 17:57:49 crc kubenswrapper[4868]: I1201 17:57:49.172781 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:57:49 crc kubenswrapper[4868]: E1201 17:57:49.174234 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:58:00 crc kubenswrapper[4868]: I1201 17:58:00.809720 4868 generic.go:334] "Generic (PLEG): container finished" podID="216fe1b0-ec90-4ee7-91e4-aa24476b39b4" containerID="6c6841afcf09b5434a25c5bbcd7c972d30ea65862b279f445f00e658538b6538" exitCode=0 Dec 01 17:58:00 crc kubenswrapper[4868]: I1201 17:58:00.809803 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" event={"ID":"216fe1b0-ec90-4ee7-91e4-aa24476b39b4","Type":"ContainerDied","Data":"6c6841afcf09b5434a25c5bbcd7c972d30ea65862b279f445f00e658538b6538"} Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.237659 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.369442 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ovn-combined-ca-bundle\") pod \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.369542 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ssh-key\") pod \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.369564 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ovncontroller-config-0\") pod \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.369606 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-inventory\") pod \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.369635 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzwnb\" (UniqueName: \"kubernetes.io/projected/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-kube-api-access-kzwnb\") pod \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\" (UID: \"216fe1b0-ec90-4ee7-91e4-aa24476b39b4\") " Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.376118 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "216fe1b0-ec90-4ee7-91e4-aa24476b39b4" (UID: "216fe1b0-ec90-4ee7-91e4-aa24476b39b4"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.376179 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-kube-api-access-kzwnb" (OuterVolumeSpecName: "kube-api-access-kzwnb") pod "216fe1b0-ec90-4ee7-91e4-aa24476b39b4" (UID: "216fe1b0-ec90-4ee7-91e4-aa24476b39b4"). InnerVolumeSpecName "kube-api-access-kzwnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.395650 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "216fe1b0-ec90-4ee7-91e4-aa24476b39b4" (UID: "216fe1b0-ec90-4ee7-91e4-aa24476b39b4"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.397659 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "216fe1b0-ec90-4ee7-91e4-aa24476b39b4" (UID: "216fe1b0-ec90-4ee7-91e4-aa24476b39b4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.398414 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-inventory" (OuterVolumeSpecName: "inventory") pod "216fe1b0-ec90-4ee7-91e4-aa24476b39b4" (UID: "216fe1b0-ec90-4ee7-91e4-aa24476b39b4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.471712 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.471982 4868 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.471996 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.472006 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzwnb\" (UniqueName: \"kubernetes.io/projected/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-kube-api-access-kzwnb\") on node \"crc\" DevicePath \"\"" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.472016 4868 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/216fe1b0-ec90-4ee7-91e4-aa24476b39b4-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.829921 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" event={"ID":"216fe1b0-ec90-4ee7-91e4-aa24476b39b4","Type":"ContainerDied","Data":"9f697d314c6af4badc76ddde68481627d1178c986adc919a4c56c0ec424e04f3"} Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.829983 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f697d314c6af4badc76ddde68481627d1178c986adc919a4c56c0ec424e04f3" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.830038 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nmn4w" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.907681 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6"] Dec 01 17:58:02 crc kubenswrapper[4868]: E1201 17:58:02.908164 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="216fe1b0-ec90-4ee7-91e4-aa24476b39b4" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.908187 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="216fe1b0-ec90-4ee7-91e4-aa24476b39b4" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 17:58:02 crc kubenswrapper[4868]: E1201 17:58:02.908206 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6549a036-6568-40dd-b26d-f1b91357e550" containerName="extract-utilities" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.908216 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="6549a036-6568-40dd-b26d-f1b91357e550" containerName="extract-utilities" Dec 01 17:58:02 crc kubenswrapper[4868]: E1201 17:58:02.908242 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6549a036-6568-40dd-b26d-f1b91357e550" containerName="registry-server" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.908249 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="6549a036-6568-40dd-b26d-f1b91357e550" containerName="registry-server" Dec 01 17:58:02 crc kubenswrapper[4868]: E1201 17:58:02.908273 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6549a036-6568-40dd-b26d-f1b91357e550" containerName="extract-content" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.908278 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="6549a036-6568-40dd-b26d-f1b91357e550" containerName="extract-content" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.908444 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="216fe1b0-ec90-4ee7-91e4-aa24476b39b4" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.908456 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="6549a036-6568-40dd-b26d-f1b91357e550" containerName="registry-server" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.909230 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.910996 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.911456 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.911611 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.911724 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.912292 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.914139 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 17:58:02 crc kubenswrapper[4868]: I1201 17:58:02.919451 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6"] Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.083166 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nb8t\" (UniqueName: \"kubernetes.io/projected/6201a05d-611d-4220-b7ae-1b0bf860b4c0-kube-api-access-9nb8t\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.083509 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.083691 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.083822 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.084115 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.084282 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.185871 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.186846 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.186890 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.187026 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.187057 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.187091 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nb8t\" (UniqueName: \"kubernetes.io/projected/6201a05d-611d-4220-b7ae-1b0bf860b4c0-kube-api-access-9nb8t\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.192295 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.192420 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.192663 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.195120 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.197197 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.203724 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nb8t\" (UniqueName: \"kubernetes.io/projected/6201a05d-611d-4220-b7ae-1b0bf860b4c0-kube-api-access-9nb8t\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.235616 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.738567 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6"] Dec 01 17:58:03 crc kubenswrapper[4868]: I1201 17:58:03.838831 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" event={"ID":"6201a05d-611d-4220-b7ae-1b0bf860b4c0","Type":"ContainerStarted","Data":"6d07a532e0c6d6c497ff01d7bfef551ecaa1426bf8da61e8565e147ecffa926f"} Dec 01 17:58:04 crc kubenswrapper[4868]: I1201 17:58:04.171707 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:58:04 crc kubenswrapper[4868]: E1201 17:58:04.172043 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:58:04 crc kubenswrapper[4868]: I1201 17:58:04.851707 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" event={"ID":"6201a05d-611d-4220-b7ae-1b0bf860b4c0","Type":"ContainerStarted","Data":"9d15bd51100afa16058a7ed085bbeb2fd4761eda3753fd72c742d209d560b468"} Dec 01 17:58:04 crc kubenswrapper[4868]: I1201 17:58:04.888134 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" podStartSLOduration=2.315242202 podStartE2EDuration="2.88810733s" podCreationTimestamp="2025-12-01 17:58:02 +0000 UTC" firstStartedPulling="2025-12-01 17:58:03.743375941 +0000 UTC m=+1956.114486352" lastFinishedPulling="2025-12-01 17:58:04.316241069 +0000 UTC m=+1956.687351480" observedRunningTime="2025-12-01 17:58:04.87301884 +0000 UTC m=+1957.244129251" watchObservedRunningTime="2025-12-01 17:58:04.88810733 +0000 UTC m=+1957.259217741" Dec 01 17:58:18 crc kubenswrapper[4868]: I1201 17:58:18.180915 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:58:18 crc kubenswrapper[4868]: E1201 17:58:18.183199 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 17:58:29 crc kubenswrapper[4868]: I1201 17:58:29.172064 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 17:58:30 crc kubenswrapper[4868]: I1201 17:58:30.136078 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"95b0a3a129d931e7326436a081957bc1421283eca7e6144750f7284aee7a3835"} Dec 01 17:58:50 crc kubenswrapper[4868]: I1201 17:58:50.335205 4868 generic.go:334] "Generic (PLEG): container finished" podID="6201a05d-611d-4220-b7ae-1b0bf860b4c0" containerID="9d15bd51100afa16058a7ed085bbeb2fd4761eda3753fd72c742d209d560b468" exitCode=0 Dec 01 17:58:50 crc kubenswrapper[4868]: I1201 17:58:50.335279 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" event={"ID":"6201a05d-611d-4220-b7ae-1b0bf860b4c0","Type":"ContainerDied","Data":"9d15bd51100afa16058a7ed085bbeb2fd4761eda3753fd72c742d209d560b468"} Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.752428 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.772561 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-ssh-key\") pod \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.772809 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.772857 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nb8t\" (UniqueName: \"kubernetes.io/projected/6201a05d-611d-4220-b7ae-1b0bf860b4c0-kube-api-access-9nb8t\") pod \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.772972 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-neutron-metadata-combined-ca-bundle\") pod \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.773060 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-inventory\") pod \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.773181 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-nova-metadata-neutron-config-0\") pod \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\" (UID: \"6201a05d-611d-4220-b7ae-1b0bf860b4c0\") " Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.780169 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "6201a05d-611d-4220-b7ae-1b0bf860b4c0" (UID: "6201a05d-611d-4220-b7ae-1b0bf860b4c0"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.784700 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6201a05d-611d-4220-b7ae-1b0bf860b4c0-kube-api-access-9nb8t" (OuterVolumeSpecName: "kube-api-access-9nb8t") pod "6201a05d-611d-4220-b7ae-1b0bf860b4c0" (UID: "6201a05d-611d-4220-b7ae-1b0bf860b4c0"). InnerVolumeSpecName "kube-api-access-9nb8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.806465 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6201a05d-611d-4220-b7ae-1b0bf860b4c0" (UID: "6201a05d-611d-4220-b7ae-1b0bf860b4c0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.808923 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "6201a05d-611d-4220-b7ae-1b0bf860b4c0" (UID: "6201a05d-611d-4220-b7ae-1b0bf860b4c0"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.818648 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "6201a05d-611d-4220-b7ae-1b0bf860b4c0" (UID: "6201a05d-611d-4220-b7ae-1b0bf860b4c0"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.819385 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-inventory" (OuterVolumeSpecName: "inventory") pod "6201a05d-611d-4220-b7ae-1b0bf860b4c0" (UID: "6201a05d-611d-4220-b7ae-1b0bf860b4c0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.875591 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nb8t\" (UniqueName: \"kubernetes.io/projected/6201a05d-611d-4220-b7ae-1b0bf860b4c0-kube-api-access-9nb8t\") on node \"crc\" DevicePath \"\"" Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.875633 4868 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.875645 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.875655 4868 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.875664 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:58:51 crc kubenswrapper[4868]: I1201 17:58:51.875674 4868 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6201a05d-611d-4220-b7ae-1b0bf860b4c0-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.356974 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" event={"ID":"6201a05d-611d-4220-b7ae-1b0bf860b4c0","Type":"ContainerDied","Data":"6d07a532e0c6d6c497ff01d7bfef551ecaa1426bf8da61e8565e147ecffa926f"} Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.357015 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d07a532e0c6d6c497ff01d7bfef551ecaa1426bf8da61e8565e147ecffa926f" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.357091 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.472840 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs"] Dec 01 17:58:52 crc kubenswrapper[4868]: E1201 17:58:52.473568 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6201a05d-611d-4220-b7ae-1b0bf860b4c0" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.473650 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="6201a05d-611d-4220-b7ae-1b0bf860b4c0" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.473887 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="6201a05d-611d-4220-b7ae-1b0bf860b4c0" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.474696 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.482047 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.482337 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.482489 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.486084 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs"] Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.487362 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.487451 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.487484 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkfrl\" (UniqueName: \"kubernetes.io/projected/944aa8d9-9aa5-466a-82fd-9768a8b970cb-kube-api-access-hkfrl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.487513 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.487529 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.526710 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.528138 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.590628 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.590769 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.590808 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkfrl\" (UniqueName: \"kubernetes.io/projected/944aa8d9-9aa5-466a-82fd-9768a8b970cb-kube-api-access-hkfrl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.590846 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.590870 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.594927 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.596544 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.597505 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.607657 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.608416 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkfrl\" (UniqueName: \"kubernetes.io/projected/944aa8d9-9aa5-466a-82fd-9768a8b970cb-kube-api-access-hkfrl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:52 crc kubenswrapper[4868]: I1201 17:58:52.837608 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 17:58:53 crc kubenswrapper[4868]: I1201 17:58:53.323343 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs"] Dec 01 17:58:53 crc kubenswrapper[4868]: I1201 17:58:53.367779 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" event={"ID":"944aa8d9-9aa5-466a-82fd-9768a8b970cb","Type":"ContainerStarted","Data":"b2351a5ee6f05528c2017db48e8ec349058c583f5d1a94993d57f05827a36256"} Dec 01 17:58:54 crc kubenswrapper[4868]: I1201 17:58:54.377522 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" event={"ID":"944aa8d9-9aa5-466a-82fd-9768a8b970cb","Type":"ContainerStarted","Data":"734414eeaf22665634fee0caf486def154a10fa23059952ab26d09eede4e809e"} Dec 01 17:58:54 crc kubenswrapper[4868]: I1201 17:58:54.402362 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" podStartSLOduration=1.739855709 podStartE2EDuration="2.40233571s" podCreationTimestamp="2025-12-01 17:58:52 +0000 UTC" firstStartedPulling="2025-12-01 17:58:53.325580355 +0000 UTC m=+2005.696690766" lastFinishedPulling="2025-12-01 17:58:53.988060356 +0000 UTC m=+2006.359170767" observedRunningTime="2025-12-01 17:58:54.391012662 +0000 UTC m=+2006.762123093" watchObservedRunningTime="2025-12-01 17:58:54.40233571 +0000 UTC m=+2006.773446121" Dec 01 18:00:00 crc kubenswrapper[4868]: I1201 18:00:00.193177 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z"] Dec 01 18:00:00 crc kubenswrapper[4868]: I1201 18:00:00.218916 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z"] Dec 01 18:00:00 crc kubenswrapper[4868]: I1201 18:00:00.220890 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" Dec 01 18:00:00 crc kubenswrapper[4868]: I1201 18:00:00.226268 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 18:00:00 crc kubenswrapper[4868]: I1201 18:00:00.226580 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 18:00:00 crc kubenswrapper[4868]: I1201 18:00:00.357575 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbn9h\" (UniqueName: \"kubernetes.io/projected/afdfce94-357c-4abf-8d97-0bbbb2999412-kube-api-access-fbn9h\") pod \"collect-profiles-29410200-hcg2z\" (UID: \"afdfce94-357c-4abf-8d97-0bbbb2999412\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" Dec 01 18:00:00 crc kubenswrapper[4868]: I1201 18:00:00.357677 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/afdfce94-357c-4abf-8d97-0bbbb2999412-secret-volume\") pod \"collect-profiles-29410200-hcg2z\" (UID: \"afdfce94-357c-4abf-8d97-0bbbb2999412\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" Dec 01 18:00:00 crc kubenswrapper[4868]: I1201 18:00:00.357774 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/afdfce94-357c-4abf-8d97-0bbbb2999412-config-volume\") pod \"collect-profiles-29410200-hcg2z\" (UID: \"afdfce94-357c-4abf-8d97-0bbbb2999412\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" Dec 01 18:00:00 crc kubenswrapper[4868]: I1201 18:00:00.459858 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbn9h\" (UniqueName: \"kubernetes.io/projected/afdfce94-357c-4abf-8d97-0bbbb2999412-kube-api-access-fbn9h\") pod \"collect-profiles-29410200-hcg2z\" (UID: \"afdfce94-357c-4abf-8d97-0bbbb2999412\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" Dec 01 18:00:00 crc kubenswrapper[4868]: I1201 18:00:00.459930 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/afdfce94-357c-4abf-8d97-0bbbb2999412-secret-volume\") pod \"collect-profiles-29410200-hcg2z\" (UID: \"afdfce94-357c-4abf-8d97-0bbbb2999412\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" Dec 01 18:00:00 crc kubenswrapper[4868]: I1201 18:00:00.460068 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/afdfce94-357c-4abf-8d97-0bbbb2999412-config-volume\") pod \"collect-profiles-29410200-hcg2z\" (UID: \"afdfce94-357c-4abf-8d97-0bbbb2999412\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" Dec 01 18:00:00 crc kubenswrapper[4868]: I1201 18:00:00.460967 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/afdfce94-357c-4abf-8d97-0bbbb2999412-config-volume\") pod \"collect-profiles-29410200-hcg2z\" (UID: \"afdfce94-357c-4abf-8d97-0bbbb2999412\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" Dec 01 18:00:00 crc kubenswrapper[4868]: I1201 18:00:00.466248 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/afdfce94-357c-4abf-8d97-0bbbb2999412-secret-volume\") pod \"collect-profiles-29410200-hcg2z\" (UID: \"afdfce94-357c-4abf-8d97-0bbbb2999412\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" Dec 01 18:00:00 crc kubenswrapper[4868]: I1201 18:00:00.477562 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbn9h\" (UniqueName: \"kubernetes.io/projected/afdfce94-357c-4abf-8d97-0bbbb2999412-kube-api-access-fbn9h\") pod \"collect-profiles-29410200-hcg2z\" (UID: \"afdfce94-357c-4abf-8d97-0bbbb2999412\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" Dec 01 18:00:00 crc kubenswrapper[4868]: I1201 18:00:00.555361 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" Dec 01 18:00:01 crc kubenswrapper[4868]: I1201 18:00:01.004519 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z"] Dec 01 18:00:01 crc kubenswrapper[4868]: I1201 18:00:01.997653 4868 generic.go:334] "Generic (PLEG): container finished" podID="afdfce94-357c-4abf-8d97-0bbbb2999412" containerID="6ba23862bb2c670552a8970fec23e3eeff913e45425943a1767de319223b5c98" exitCode=0 Dec 01 18:00:01 crc kubenswrapper[4868]: I1201 18:00:01.997721 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" event={"ID":"afdfce94-357c-4abf-8d97-0bbbb2999412","Type":"ContainerDied","Data":"6ba23862bb2c670552a8970fec23e3eeff913e45425943a1767de319223b5c98"} Dec 01 18:00:01 crc kubenswrapper[4868]: I1201 18:00:01.998312 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" event={"ID":"afdfce94-357c-4abf-8d97-0bbbb2999412","Type":"ContainerStarted","Data":"e147b3751c8b1647c2e0e28b0ef139a47fc381774a941c2e478801d6d9dca2eb"} Dec 01 18:00:03 crc kubenswrapper[4868]: I1201 18:00:03.342686 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" Dec 01 18:00:03 crc kubenswrapper[4868]: I1201 18:00:03.421341 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/afdfce94-357c-4abf-8d97-0bbbb2999412-secret-volume\") pod \"afdfce94-357c-4abf-8d97-0bbbb2999412\" (UID: \"afdfce94-357c-4abf-8d97-0bbbb2999412\") " Dec 01 18:00:03 crc kubenswrapper[4868]: I1201 18:00:03.421481 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/afdfce94-357c-4abf-8d97-0bbbb2999412-config-volume\") pod \"afdfce94-357c-4abf-8d97-0bbbb2999412\" (UID: \"afdfce94-357c-4abf-8d97-0bbbb2999412\") " Dec 01 18:00:03 crc kubenswrapper[4868]: I1201 18:00:03.421575 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbn9h\" (UniqueName: \"kubernetes.io/projected/afdfce94-357c-4abf-8d97-0bbbb2999412-kube-api-access-fbn9h\") pod \"afdfce94-357c-4abf-8d97-0bbbb2999412\" (UID: \"afdfce94-357c-4abf-8d97-0bbbb2999412\") " Dec 01 18:00:03 crc kubenswrapper[4868]: I1201 18:00:03.423613 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afdfce94-357c-4abf-8d97-0bbbb2999412-config-volume" (OuterVolumeSpecName: "config-volume") pod "afdfce94-357c-4abf-8d97-0bbbb2999412" (UID: "afdfce94-357c-4abf-8d97-0bbbb2999412"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 18:00:03 crc kubenswrapper[4868]: I1201 18:00:03.436148 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdfce94-357c-4abf-8d97-0bbbb2999412-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "afdfce94-357c-4abf-8d97-0bbbb2999412" (UID: "afdfce94-357c-4abf-8d97-0bbbb2999412"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:00:03 crc kubenswrapper[4868]: I1201 18:00:03.436196 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afdfce94-357c-4abf-8d97-0bbbb2999412-kube-api-access-fbn9h" (OuterVolumeSpecName: "kube-api-access-fbn9h") pod "afdfce94-357c-4abf-8d97-0bbbb2999412" (UID: "afdfce94-357c-4abf-8d97-0bbbb2999412"). InnerVolumeSpecName "kube-api-access-fbn9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:00:03 crc kubenswrapper[4868]: I1201 18:00:03.524901 4868 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/afdfce94-357c-4abf-8d97-0bbbb2999412-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 18:00:03 crc kubenswrapper[4868]: I1201 18:00:03.525424 4868 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/afdfce94-357c-4abf-8d97-0bbbb2999412-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 18:00:03 crc kubenswrapper[4868]: I1201 18:00:03.525438 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbn9h\" (UniqueName: \"kubernetes.io/projected/afdfce94-357c-4abf-8d97-0bbbb2999412-kube-api-access-fbn9h\") on node \"crc\" DevicePath \"\"" Dec 01 18:00:04 crc kubenswrapper[4868]: I1201 18:00:04.022538 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" event={"ID":"afdfce94-357c-4abf-8d97-0bbbb2999412","Type":"ContainerDied","Data":"e147b3751c8b1647c2e0e28b0ef139a47fc381774a941c2e478801d6d9dca2eb"} Dec 01 18:00:04 crc kubenswrapper[4868]: I1201 18:00:04.022583 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e147b3751c8b1647c2e0e28b0ef139a47fc381774a941c2e478801d6d9dca2eb" Dec 01 18:00:04 crc kubenswrapper[4868]: I1201 18:00:04.022659 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410200-hcg2z" Dec 01 18:00:04 crc kubenswrapper[4868]: I1201 18:00:04.420576 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m"] Dec 01 18:00:04 crc kubenswrapper[4868]: I1201 18:00:04.429255 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410155-bvw5m"] Dec 01 18:00:06 crc kubenswrapper[4868]: I1201 18:00:06.184129 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="962e4f14-b280-4ffa-bde0-6cc4b0d86144" path="/var/lib/kubelet/pods/962e4f14-b280-4ffa-bde0-6cc4b0d86144/volumes" Dec 01 18:00:41 crc kubenswrapper[4868]: I1201 18:00:41.308335 4868 scope.go:117] "RemoveContainer" containerID="83cbd781a24db21510be592699a361e90e624e963509bd92bbef568a0b050db1" Dec 01 18:00:55 crc kubenswrapper[4868]: I1201 18:00:55.904473 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:00:55 crc kubenswrapper[4868]: I1201 18:00:55.905056 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.152934 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29410201-xgqzm"] Dec 01 18:01:00 crc kubenswrapper[4868]: E1201 18:01:00.153875 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afdfce94-357c-4abf-8d97-0bbbb2999412" containerName="collect-profiles" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.153894 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="afdfce94-357c-4abf-8d97-0bbbb2999412" containerName="collect-profiles" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.154157 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="afdfce94-357c-4abf-8d97-0bbbb2999412" containerName="collect-profiles" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.154862 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.183005 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29410201-xgqzm"] Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.285064 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-config-data\") pod \"keystone-cron-29410201-xgqzm\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.285127 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-fernet-keys\") pod \"keystone-cron-29410201-xgqzm\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.285872 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8862\" (UniqueName: \"kubernetes.io/projected/42f06b5b-cb48-40da-8a82-f3af58170592-kube-api-access-j8862\") pod \"keystone-cron-29410201-xgqzm\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.285904 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-combined-ca-bundle\") pod \"keystone-cron-29410201-xgqzm\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.387681 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8862\" (UniqueName: \"kubernetes.io/projected/42f06b5b-cb48-40da-8a82-f3af58170592-kube-api-access-j8862\") pod \"keystone-cron-29410201-xgqzm\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.387732 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-combined-ca-bundle\") pod \"keystone-cron-29410201-xgqzm\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.387840 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-config-data\") pod \"keystone-cron-29410201-xgqzm\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.387858 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-fernet-keys\") pod \"keystone-cron-29410201-xgqzm\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.398957 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-fernet-keys\") pod \"keystone-cron-29410201-xgqzm\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.399970 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-config-data\") pod \"keystone-cron-29410201-xgqzm\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.401445 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-combined-ca-bundle\") pod \"keystone-cron-29410201-xgqzm\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.405000 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8862\" (UniqueName: \"kubernetes.io/projected/42f06b5b-cb48-40da-8a82-f3af58170592-kube-api-access-j8862\") pod \"keystone-cron-29410201-xgqzm\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.477283 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:00 crc kubenswrapper[4868]: I1201 18:01:00.996000 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29410201-xgqzm"] Dec 01 18:01:01 crc kubenswrapper[4868]: I1201 18:01:01.678701 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410201-xgqzm" event={"ID":"42f06b5b-cb48-40da-8a82-f3af58170592","Type":"ContainerStarted","Data":"0288f2dc7e556dff3ce07588bdfbbb6090a5d26d2d446b18c07825c637ccfd02"} Dec 01 18:01:01 crc kubenswrapper[4868]: I1201 18:01:01.679733 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410201-xgqzm" event={"ID":"42f06b5b-cb48-40da-8a82-f3af58170592","Type":"ContainerStarted","Data":"a01a78565fe0ca089480c4ef283b0e3f34cc7545f3b3b477f0d4f433e1529a28"} Dec 01 18:01:03 crc kubenswrapper[4868]: I1201 18:01:03.701104 4868 generic.go:334] "Generic (PLEG): container finished" podID="42f06b5b-cb48-40da-8a82-f3af58170592" containerID="0288f2dc7e556dff3ce07588bdfbbb6090a5d26d2d446b18c07825c637ccfd02" exitCode=0 Dec 01 18:01:03 crc kubenswrapper[4868]: I1201 18:01:03.701203 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410201-xgqzm" event={"ID":"42f06b5b-cb48-40da-8a82-f3af58170592","Type":"ContainerDied","Data":"0288f2dc7e556dff3ce07588bdfbbb6090a5d26d2d446b18c07825c637ccfd02"} Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.156316 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.292848 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-fernet-keys\") pod \"42f06b5b-cb48-40da-8a82-f3af58170592\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.292958 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-config-data\") pod \"42f06b5b-cb48-40da-8a82-f3af58170592\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.292978 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-combined-ca-bundle\") pod \"42f06b5b-cb48-40da-8a82-f3af58170592\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.293063 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8862\" (UniqueName: \"kubernetes.io/projected/42f06b5b-cb48-40da-8a82-f3af58170592-kube-api-access-j8862\") pod \"42f06b5b-cb48-40da-8a82-f3af58170592\" (UID: \"42f06b5b-cb48-40da-8a82-f3af58170592\") " Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.299649 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "42f06b5b-cb48-40da-8a82-f3af58170592" (UID: "42f06b5b-cb48-40da-8a82-f3af58170592"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.300531 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42f06b5b-cb48-40da-8a82-f3af58170592-kube-api-access-j8862" (OuterVolumeSpecName: "kube-api-access-j8862") pod "42f06b5b-cb48-40da-8a82-f3af58170592" (UID: "42f06b5b-cb48-40da-8a82-f3af58170592"). InnerVolumeSpecName "kube-api-access-j8862". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.322348 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42f06b5b-cb48-40da-8a82-f3af58170592" (UID: "42f06b5b-cb48-40da-8a82-f3af58170592"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.352869 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-config-data" (OuterVolumeSpecName: "config-data") pod "42f06b5b-cb48-40da-8a82-f3af58170592" (UID: "42f06b5b-cb48-40da-8a82-f3af58170592"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.396418 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8862\" (UniqueName: \"kubernetes.io/projected/42f06b5b-cb48-40da-8a82-f3af58170592-kube-api-access-j8862\") on node \"crc\" DevicePath \"\"" Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.396455 4868 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.396465 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.396478 4868 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42f06b5b-cb48-40da-8a82-f3af58170592-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.727588 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410201-xgqzm" event={"ID":"42f06b5b-cb48-40da-8a82-f3af58170592","Type":"ContainerDied","Data":"a01a78565fe0ca089480c4ef283b0e3f34cc7545f3b3b477f0d4f433e1529a28"} Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.727647 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a01a78565fe0ca089480c4ef283b0e3f34cc7545f3b3b477f0d4f433e1529a28" Dec 01 18:01:05 crc kubenswrapper[4868]: I1201 18:01:05.727700 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410201-xgqzm" Dec 01 18:01:25 crc kubenswrapper[4868]: I1201 18:01:25.904917 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:01:25 crc kubenswrapper[4868]: I1201 18:01:25.905506 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:01:44 crc kubenswrapper[4868]: I1201 18:01:44.477745 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rqplr"] Dec 01 18:01:44 crc kubenswrapper[4868]: E1201 18:01:44.478678 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f06b5b-cb48-40da-8a82-f3af58170592" containerName="keystone-cron" Dec 01 18:01:44 crc kubenswrapper[4868]: I1201 18:01:44.478692 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f06b5b-cb48-40da-8a82-f3af58170592" containerName="keystone-cron" Dec 01 18:01:44 crc kubenswrapper[4868]: I1201 18:01:44.478889 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="42f06b5b-cb48-40da-8a82-f3af58170592" containerName="keystone-cron" Dec 01 18:01:44 crc kubenswrapper[4868]: I1201 18:01:44.480341 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:44 crc kubenswrapper[4868]: I1201 18:01:44.498822 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rqplr"] Dec 01 18:01:44 crc kubenswrapper[4868]: I1201 18:01:44.581130 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b880bd-ac45-4fee-9a95-9fbd38285f18-utilities\") pod \"redhat-marketplace-rqplr\" (UID: \"b4b880bd-ac45-4fee-9a95-9fbd38285f18\") " pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:44 crc kubenswrapper[4868]: I1201 18:01:44.581245 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b880bd-ac45-4fee-9a95-9fbd38285f18-catalog-content\") pod \"redhat-marketplace-rqplr\" (UID: \"b4b880bd-ac45-4fee-9a95-9fbd38285f18\") " pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:44 crc kubenswrapper[4868]: I1201 18:01:44.581354 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5l5t\" (UniqueName: \"kubernetes.io/projected/b4b880bd-ac45-4fee-9a95-9fbd38285f18-kube-api-access-l5l5t\") pod \"redhat-marketplace-rqplr\" (UID: \"b4b880bd-ac45-4fee-9a95-9fbd38285f18\") " pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:44 crc kubenswrapper[4868]: I1201 18:01:44.683671 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b880bd-ac45-4fee-9a95-9fbd38285f18-catalog-content\") pod \"redhat-marketplace-rqplr\" (UID: \"b4b880bd-ac45-4fee-9a95-9fbd38285f18\") " pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:44 crc kubenswrapper[4868]: I1201 18:01:44.683780 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5l5t\" (UniqueName: \"kubernetes.io/projected/b4b880bd-ac45-4fee-9a95-9fbd38285f18-kube-api-access-l5l5t\") pod \"redhat-marketplace-rqplr\" (UID: \"b4b880bd-ac45-4fee-9a95-9fbd38285f18\") " pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:44 crc kubenswrapper[4868]: I1201 18:01:44.684004 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b880bd-ac45-4fee-9a95-9fbd38285f18-utilities\") pod \"redhat-marketplace-rqplr\" (UID: \"b4b880bd-ac45-4fee-9a95-9fbd38285f18\") " pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:44 crc kubenswrapper[4868]: I1201 18:01:44.684603 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b880bd-ac45-4fee-9a95-9fbd38285f18-utilities\") pod \"redhat-marketplace-rqplr\" (UID: \"b4b880bd-ac45-4fee-9a95-9fbd38285f18\") " pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:44 crc kubenswrapper[4868]: I1201 18:01:44.684665 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b880bd-ac45-4fee-9a95-9fbd38285f18-catalog-content\") pod \"redhat-marketplace-rqplr\" (UID: \"b4b880bd-ac45-4fee-9a95-9fbd38285f18\") " pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:44 crc kubenswrapper[4868]: I1201 18:01:44.706108 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5l5t\" (UniqueName: \"kubernetes.io/projected/b4b880bd-ac45-4fee-9a95-9fbd38285f18-kube-api-access-l5l5t\") pod \"redhat-marketplace-rqplr\" (UID: \"b4b880bd-ac45-4fee-9a95-9fbd38285f18\") " pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:44 crc kubenswrapper[4868]: I1201 18:01:44.804054 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:45 crc kubenswrapper[4868]: I1201 18:01:45.377576 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rqplr"] Dec 01 18:01:45 crc kubenswrapper[4868]: I1201 18:01:45.597188 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqplr" event={"ID":"b4b880bd-ac45-4fee-9a95-9fbd38285f18","Type":"ContainerStarted","Data":"faecb16733265b91d85f7802d72df005f4d63febd0c4b2e6c0d53d92f573aaed"} Dec 01 18:01:46 crc kubenswrapper[4868]: I1201 18:01:46.608544 4868 generic.go:334] "Generic (PLEG): container finished" podID="b4b880bd-ac45-4fee-9a95-9fbd38285f18" containerID="f793420bf771fd114eb784a99d6f5e59ad9f1ef910839d231cf3ead161e447cc" exitCode=0 Dec 01 18:01:46 crc kubenswrapper[4868]: I1201 18:01:46.608602 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqplr" event={"ID":"b4b880bd-ac45-4fee-9a95-9fbd38285f18","Type":"ContainerDied","Data":"f793420bf771fd114eb784a99d6f5e59ad9f1ef910839d231cf3ead161e447cc"} Dec 01 18:01:46 crc kubenswrapper[4868]: I1201 18:01:46.610999 4868 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 18:01:47 crc kubenswrapper[4868]: I1201 18:01:47.620460 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqplr" event={"ID":"b4b880bd-ac45-4fee-9a95-9fbd38285f18","Type":"ContainerStarted","Data":"de84f80b88ca0caea9369586fba32540faa1ff3055b8e80b9f7c2c40b6fdfcfa"} Dec 01 18:01:48 crc kubenswrapper[4868]: I1201 18:01:48.284375 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fpxc7"] Dec 01 18:01:48 crc kubenswrapper[4868]: I1201 18:01:48.301381 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fpxc7"] Dec 01 18:01:48 crc kubenswrapper[4868]: I1201 18:01:48.301546 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fpxc7" Dec 01 18:01:48 crc kubenswrapper[4868]: I1201 18:01:48.478310 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls8nr\" (UniqueName: \"kubernetes.io/projected/ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a-kube-api-access-ls8nr\") pod \"community-operators-fpxc7\" (UID: \"ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a\") " pod="openshift-marketplace/community-operators-fpxc7" Dec 01 18:01:48 crc kubenswrapper[4868]: I1201 18:01:48.478997 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a-catalog-content\") pod \"community-operators-fpxc7\" (UID: \"ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a\") " pod="openshift-marketplace/community-operators-fpxc7" Dec 01 18:01:48 crc kubenswrapper[4868]: I1201 18:01:48.479196 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a-utilities\") pod \"community-operators-fpxc7\" (UID: \"ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a\") " pod="openshift-marketplace/community-operators-fpxc7" Dec 01 18:01:48 crc kubenswrapper[4868]: I1201 18:01:48.588202 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a-catalog-content\") pod \"community-operators-fpxc7\" (UID: \"ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a\") " pod="openshift-marketplace/community-operators-fpxc7" Dec 01 18:01:48 crc kubenswrapper[4868]: I1201 18:01:48.588317 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a-utilities\") pod \"community-operators-fpxc7\" (UID: \"ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a\") " pod="openshift-marketplace/community-operators-fpxc7" Dec 01 18:01:48 crc kubenswrapper[4868]: I1201 18:01:48.588419 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls8nr\" (UniqueName: \"kubernetes.io/projected/ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a-kube-api-access-ls8nr\") pod \"community-operators-fpxc7\" (UID: \"ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a\") " pod="openshift-marketplace/community-operators-fpxc7" Dec 01 18:01:48 crc kubenswrapper[4868]: I1201 18:01:48.588686 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a-catalog-content\") pod \"community-operators-fpxc7\" (UID: \"ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a\") " pod="openshift-marketplace/community-operators-fpxc7" Dec 01 18:01:48 crc kubenswrapper[4868]: I1201 18:01:48.588907 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a-utilities\") pod \"community-operators-fpxc7\" (UID: \"ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a\") " pod="openshift-marketplace/community-operators-fpxc7" Dec 01 18:01:48 crc kubenswrapper[4868]: I1201 18:01:48.626262 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls8nr\" (UniqueName: \"kubernetes.io/projected/ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a-kube-api-access-ls8nr\") pod \"community-operators-fpxc7\" (UID: \"ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a\") " pod="openshift-marketplace/community-operators-fpxc7" Dec 01 18:01:48 crc kubenswrapper[4868]: I1201 18:01:48.626768 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fpxc7" Dec 01 18:01:48 crc kubenswrapper[4868]: I1201 18:01:48.636791 4868 generic.go:334] "Generic (PLEG): container finished" podID="b4b880bd-ac45-4fee-9a95-9fbd38285f18" containerID="de84f80b88ca0caea9369586fba32540faa1ff3055b8e80b9f7c2c40b6fdfcfa" exitCode=0 Dec 01 18:01:48 crc kubenswrapper[4868]: I1201 18:01:48.636844 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqplr" event={"ID":"b4b880bd-ac45-4fee-9a95-9fbd38285f18","Type":"ContainerDied","Data":"de84f80b88ca0caea9369586fba32540faa1ff3055b8e80b9f7c2c40b6fdfcfa"} Dec 01 18:01:49 crc kubenswrapper[4868]: I1201 18:01:49.185336 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fpxc7"] Dec 01 18:01:49 crc kubenswrapper[4868]: I1201 18:01:49.647401 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqplr" event={"ID":"b4b880bd-ac45-4fee-9a95-9fbd38285f18","Type":"ContainerStarted","Data":"36c97e10431cda529464093b168dd0c9e3eb0c01ac61b3523d6bcd711bf3536b"} Dec 01 18:01:49 crc kubenswrapper[4868]: I1201 18:01:49.649671 4868 generic.go:334] "Generic (PLEG): container finished" podID="ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a" containerID="20e8367a34bc99a0d9bba3c4fa269942937471ae10f0bcfac4d8c6c59700f39f" exitCode=0 Dec 01 18:01:49 crc kubenswrapper[4868]: I1201 18:01:49.649765 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fpxc7" event={"ID":"ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a","Type":"ContainerDied","Data":"20e8367a34bc99a0d9bba3c4fa269942937471ae10f0bcfac4d8c6c59700f39f"} Dec 01 18:01:49 crc kubenswrapper[4868]: I1201 18:01:49.649804 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fpxc7" event={"ID":"ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a","Type":"ContainerStarted","Data":"7f81711d5f86d3ca80847b71afaab0bd99b8416060611b5c9a7e3c920bd4651d"} Dec 01 18:01:49 crc kubenswrapper[4868]: I1201 18:01:49.675192 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rqplr" podStartSLOduration=3.139989394 podStartE2EDuration="5.675175823s" podCreationTimestamp="2025-12-01 18:01:44 +0000 UTC" firstStartedPulling="2025-12-01 18:01:46.610707528 +0000 UTC m=+2178.981817939" lastFinishedPulling="2025-12-01 18:01:49.145893957 +0000 UTC m=+2181.517004368" observedRunningTime="2025-12-01 18:01:49.673840087 +0000 UTC m=+2182.044950488" watchObservedRunningTime="2025-12-01 18:01:49.675175823 +0000 UTC m=+2182.046286234" Dec 01 18:01:54 crc kubenswrapper[4868]: I1201 18:01:54.805194 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:54 crc kubenswrapper[4868]: I1201 18:01:54.806149 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:54 crc kubenswrapper[4868]: I1201 18:01:54.889089 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:55 crc kubenswrapper[4868]: I1201 18:01:55.752750 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:55 crc kubenswrapper[4868]: I1201 18:01:55.812923 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rqplr"] Dec 01 18:01:55 crc kubenswrapper[4868]: I1201 18:01:55.905883 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:01:55 crc kubenswrapper[4868]: I1201 18:01:55.905966 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:01:55 crc kubenswrapper[4868]: I1201 18:01:55.906022 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 18:01:55 crc kubenswrapper[4868]: I1201 18:01:55.906797 4868 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"95b0a3a129d931e7326436a081957bc1421283eca7e6144750f7284aee7a3835"} pod="openshift-machine-config-operator/machine-config-daemon-njgpn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 18:01:55 crc kubenswrapper[4868]: I1201 18:01:55.906858 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" containerID="cri-o://95b0a3a129d931e7326436a081957bc1421283eca7e6144750f7284aee7a3835" gracePeriod=600 Dec 01 18:01:56 crc kubenswrapper[4868]: I1201 18:01:56.715902 4868 generic.go:334] "Generic (PLEG): container finished" podID="ec4e08bf-df56-4490-8e79-a4175c727405" containerID="95b0a3a129d931e7326436a081957bc1421283eca7e6144750f7284aee7a3835" exitCode=0 Dec 01 18:01:56 crc kubenswrapper[4868]: I1201 18:01:56.715979 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerDied","Data":"95b0a3a129d931e7326436a081957bc1421283eca7e6144750f7284aee7a3835"} Dec 01 18:01:56 crc kubenswrapper[4868]: I1201 18:01:56.716638 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892"} Dec 01 18:01:56 crc kubenswrapper[4868]: I1201 18:01:56.716666 4868 scope.go:117] "RemoveContainer" containerID="5436a86615e5de422781e1104ae536628e8dab2d7c008cdee851e1605114fded" Dec 01 18:01:56 crc kubenswrapper[4868]: I1201 18:01:56.719603 4868 generic.go:334] "Generic (PLEG): container finished" podID="ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a" containerID="ca2b0ff954877b0aaf8128babff4f3a29694291b5539cdcac79be95c6bea5c0d" exitCode=0 Dec 01 18:01:56 crc kubenswrapper[4868]: I1201 18:01:56.719853 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fpxc7" event={"ID":"ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a","Type":"ContainerDied","Data":"ca2b0ff954877b0aaf8128babff4f3a29694291b5539cdcac79be95c6bea5c0d"} Dec 01 18:01:57 crc kubenswrapper[4868]: I1201 18:01:57.733714 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fpxc7" event={"ID":"ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a","Type":"ContainerStarted","Data":"324929b1b76e9f8aa51b01fcc52cbafb6566fbc003287f3df3b049820f20ab5f"} Dec 01 18:01:57 crc kubenswrapper[4868]: I1201 18:01:57.733834 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rqplr" podUID="b4b880bd-ac45-4fee-9a95-9fbd38285f18" containerName="registry-server" containerID="cri-o://36c97e10431cda529464093b168dd0c9e3eb0c01ac61b3523d6bcd711bf3536b" gracePeriod=2 Dec 01 18:01:57 crc kubenswrapper[4868]: I1201 18:01:57.758312 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fpxc7" podStartSLOduration=2.04485761 podStartE2EDuration="9.758290868s" podCreationTimestamp="2025-12-01 18:01:48 +0000 UTC" firstStartedPulling="2025-12-01 18:01:49.651338415 +0000 UTC m=+2182.022448826" lastFinishedPulling="2025-12-01 18:01:57.364771673 +0000 UTC m=+2189.735882084" observedRunningTime="2025-12-01 18:01:57.758085033 +0000 UTC m=+2190.129195444" watchObservedRunningTime="2025-12-01 18:01:57.758290868 +0000 UTC m=+2190.129401279" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.139626 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d2g2b"] Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.142517 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.158638 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d2g2b"] Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.210891 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.224596 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5gmg\" (UniqueName: \"kubernetes.io/projected/0ba0b675-4358-4b02-81f1-d6223ad87229-kube-api-access-w5gmg\") pod \"redhat-operators-d2g2b\" (UID: \"0ba0b675-4358-4b02-81f1-d6223ad87229\") " pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.224763 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ba0b675-4358-4b02-81f1-d6223ad87229-utilities\") pod \"redhat-operators-d2g2b\" (UID: \"0ba0b675-4358-4b02-81f1-d6223ad87229\") " pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.224790 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ba0b675-4358-4b02-81f1-d6223ad87229-catalog-content\") pod \"redhat-operators-d2g2b\" (UID: \"0ba0b675-4358-4b02-81f1-d6223ad87229\") " pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.326047 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5l5t\" (UniqueName: \"kubernetes.io/projected/b4b880bd-ac45-4fee-9a95-9fbd38285f18-kube-api-access-l5l5t\") pod \"b4b880bd-ac45-4fee-9a95-9fbd38285f18\" (UID: \"b4b880bd-ac45-4fee-9a95-9fbd38285f18\") " Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.326523 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b880bd-ac45-4fee-9a95-9fbd38285f18-catalog-content\") pod \"b4b880bd-ac45-4fee-9a95-9fbd38285f18\" (UID: \"b4b880bd-ac45-4fee-9a95-9fbd38285f18\") " Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.326636 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b880bd-ac45-4fee-9a95-9fbd38285f18-utilities\") pod \"b4b880bd-ac45-4fee-9a95-9fbd38285f18\" (UID: \"b4b880bd-ac45-4fee-9a95-9fbd38285f18\") " Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.327070 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ba0b675-4358-4b02-81f1-d6223ad87229-utilities\") pod \"redhat-operators-d2g2b\" (UID: \"0ba0b675-4358-4b02-81f1-d6223ad87229\") " pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.327098 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ba0b675-4358-4b02-81f1-d6223ad87229-catalog-content\") pod \"redhat-operators-d2g2b\" (UID: \"0ba0b675-4358-4b02-81f1-d6223ad87229\") " pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.327198 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5gmg\" (UniqueName: \"kubernetes.io/projected/0ba0b675-4358-4b02-81f1-d6223ad87229-kube-api-access-w5gmg\") pod \"redhat-operators-d2g2b\" (UID: \"0ba0b675-4358-4b02-81f1-d6223ad87229\") " pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.327787 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ba0b675-4358-4b02-81f1-d6223ad87229-catalog-content\") pod \"redhat-operators-d2g2b\" (UID: \"0ba0b675-4358-4b02-81f1-d6223ad87229\") " pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.327818 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ba0b675-4358-4b02-81f1-d6223ad87229-utilities\") pod \"redhat-operators-d2g2b\" (UID: \"0ba0b675-4358-4b02-81f1-d6223ad87229\") " pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.327859 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4b880bd-ac45-4fee-9a95-9fbd38285f18-utilities" (OuterVolumeSpecName: "utilities") pod "b4b880bd-ac45-4fee-9a95-9fbd38285f18" (UID: "b4b880bd-ac45-4fee-9a95-9fbd38285f18"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.337139 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4b880bd-ac45-4fee-9a95-9fbd38285f18-kube-api-access-l5l5t" (OuterVolumeSpecName: "kube-api-access-l5l5t") pod "b4b880bd-ac45-4fee-9a95-9fbd38285f18" (UID: "b4b880bd-ac45-4fee-9a95-9fbd38285f18"). InnerVolumeSpecName "kube-api-access-l5l5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.341974 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4b880bd-ac45-4fee-9a95-9fbd38285f18-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4b880bd-ac45-4fee-9a95-9fbd38285f18" (UID: "b4b880bd-ac45-4fee-9a95-9fbd38285f18"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.347875 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5gmg\" (UniqueName: \"kubernetes.io/projected/0ba0b675-4358-4b02-81f1-d6223ad87229-kube-api-access-w5gmg\") pod \"redhat-operators-d2g2b\" (UID: \"0ba0b675-4358-4b02-81f1-d6223ad87229\") " pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.429669 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5l5t\" (UniqueName: \"kubernetes.io/projected/b4b880bd-ac45-4fee-9a95-9fbd38285f18-kube-api-access-l5l5t\") on node \"crc\" DevicePath \"\"" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.429704 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b880bd-ac45-4fee-9a95-9fbd38285f18-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.429714 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b880bd-ac45-4fee-9a95-9fbd38285f18-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.522820 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.626975 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fpxc7" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.627035 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fpxc7" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.750751 4868 generic.go:334] "Generic (PLEG): container finished" podID="b4b880bd-ac45-4fee-9a95-9fbd38285f18" containerID="36c97e10431cda529464093b168dd0c9e3eb0c01ac61b3523d6bcd711bf3536b" exitCode=0 Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.751915 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqplr" event={"ID":"b4b880bd-ac45-4fee-9a95-9fbd38285f18","Type":"ContainerDied","Data":"36c97e10431cda529464093b168dd0c9e3eb0c01ac61b3523d6bcd711bf3536b"} Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.751974 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqplr" event={"ID":"b4b880bd-ac45-4fee-9a95-9fbd38285f18","Type":"ContainerDied","Data":"faecb16733265b91d85f7802d72df005f4d63febd0c4b2e6c0d53d92f573aaed"} Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.751994 4868 scope.go:117] "RemoveContainer" containerID="36c97e10431cda529464093b168dd0c9e3eb0c01ac61b3523d6bcd711bf3536b" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.752048 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rqplr" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.791069 4868 scope.go:117] "RemoveContainer" containerID="de84f80b88ca0caea9369586fba32540faa1ff3055b8e80b9f7c2c40b6fdfcfa" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.839217 4868 scope.go:117] "RemoveContainer" containerID="f793420bf771fd114eb784a99d6f5e59ad9f1ef910839d231cf3ead161e447cc" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.851256 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rqplr"] Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.872272 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rqplr"] Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.876577 4868 scope.go:117] "RemoveContainer" containerID="36c97e10431cda529464093b168dd0c9e3eb0c01ac61b3523d6bcd711bf3536b" Dec 01 18:01:58 crc kubenswrapper[4868]: E1201 18:01:58.878315 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36c97e10431cda529464093b168dd0c9e3eb0c01ac61b3523d6bcd711bf3536b\": container with ID starting with 36c97e10431cda529464093b168dd0c9e3eb0c01ac61b3523d6bcd711bf3536b not found: ID does not exist" containerID="36c97e10431cda529464093b168dd0c9e3eb0c01ac61b3523d6bcd711bf3536b" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.878373 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36c97e10431cda529464093b168dd0c9e3eb0c01ac61b3523d6bcd711bf3536b"} err="failed to get container status \"36c97e10431cda529464093b168dd0c9e3eb0c01ac61b3523d6bcd711bf3536b\": rpc error: code = NotFound desc = could not find container \"36c97e10431cda529464093b168dd0c9e3eb0c01ac61b3523d6bcd711bf3536b\": container with ID starting with 36c97e10431cda529464093b168dd0c9e3eb0c01ac61b3523d6bcd711bf3536b not found: ID does not exist" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.878404 4868 scope.go:117] "RemoveContainer" containerID="de84f80b88ca0caea9369586fba32540faa1ff3055b8e80b9f7c2c40b6fdfcfa" Dec 01 18:01:58 crc kubenswrapper[4868]: E1201 18:01:58.879634 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de84f80b88ca0caea9369586fba32540faa1ff3055b8e80b9f7c2c40b6fdfcfa\": container with ID starting with de84f80b88ca0caea9369586fba32540faa1ff3055b8e80b9f7c2c40b6fdfcfa not found: ID does not exist" containerID="de84f80b88ca0caea9369586fba32540faa1ff3055b8e80b9f7c2c40b6fdfcfa" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.879714 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de84f80b88ca0caea9369586fba32540faa1ff3055b8e80b9f7c2c40b6fdfcfa"} err="failed to get container status \"de84f80b88ca0caea9369586fba32540faa1ff3055b8e80b9f7c2c40b6fdfcfa\": rpc error: code = NotFound desc = could not find container \"de84f80b88ca0caea9369586fba32540faa1ff3055b8e80b9f7c2c40b6fdfcfa\": container with ID starting with de84f80b88ca0caea9369586fba32540faa1ff3055b8e80b9f7c2c40b6fdfcfa not found: ID does not exist" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.879752 4868 scope.go:117] "RemoveContainer" containerID="f793420bf771fd114eb784a99d6f5e59ad9f1ef910839d231cf3ead161e447cc" Dec 01 18:01:58 crc kubenswrapper[4868]: E1201 18:01:58.883347 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f793420bf771fd114eb784a99d6f5e59ad9f1ef910839d231cf3ead161e447cc\": container with ID starting with f793420bf771fd114eb784a99d6f5e59ad9f1ef910839d231cf3ead161e447cc not found: ID does not exist" containerID="f793420bf771fd114eb784a99d6f5e59ad9f1ef910839d231cf3ead161e447cc" Dec 01 18:01:58 crc kubenswrapper[4868]: I1201 18:01:58.883464 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f793420bf771fd114eb784a99d6f5e59ad9f1ef910839d231cf3ead161e447cc"} err="failed to get container status \"f793420bf771fd114eb784a99d6f5e59ad9f1ef910839d231cf3ead161e447cc\": rpc error: code = NotFound desc = could not find container \"f793420bf771fd114eb784a99d6f5e59ad9f1ef910839d231cf3ead161e447cc\": container with ID starting with f793420bf771fd114eb784a99d6f5e59ad9f1ef910839d231cf3ead161e447cc not found: ID does not exist" Dec 01 18:01:59 crc kubenswrapper[4868]: I1201 18:01:59.019669 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d2g2b"] Dec 01 18:01:59 crc kubenswrapper[4868]: I1201 18:01:59.702133 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-fpxc7" podUID="ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a" containerName="registry-server" probeResult="failure" output=< Dec 01 18:01:59 crc kubenswrapper[4868]: timeout: failed to connect service ":50051" within 1s Dec 01 18:01:59 crc kubenswrapper[4868]: > Dec 01 18:01:59 crc kubenswrapper[4868]: I1201 18:01:59.764714 4868 generic.go:334] "Generic (PLEG): container finished" podID="0ba0b675-4358-4b02-81f1-d6223ad87229" containerID="8d05ee04fb47291b0a8cce8ea89cb6a4cdc53838eaac06b96d2bc1d57ebffd17" exitCode=0 Dec 01 18:01:59 crc kubenswrapper[4868]: I1201 18:01:59.764871 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2g2b" event={"ID":"0ba0b675-4358-4b02-81f1-d6223ad87229","Type":"ContainerDied","Data":"8d05ee04fb47291b0a8cce8ea89cb6a4cdc53838eaac06b96d2bc1d57ebffd17"} Dec 01 18:01:59 crc kubenswrapper[4868]: I1201 18:01:59.765013 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2g2b" event={"ID":"0ba0b675-4358-4b02-81f1-d6223ad87229","Type":"ContainerStarted","Data":"680a0e3d439d6fae9bb51a069da57e3510fba6e296b4800a68873f6ca21c60a3"} Dec 01 18:02:00 crc kubenswrapper[4868]: I1201 18:02:00.182546 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4b880bd-ac45-4fee-9a95-9fbd38285f18" path="/var/lib/kubelet/pods/b4b880bd-ac45-4fee-9a95-9fbd38285f18/volumes" Dec 01 18:02:01 crc kubenswrapper[4868]: I1201 18:02:01.785715 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2g2b" event={"ID":"0ba0b675-4358-4b02-81f1-d6223ad87229","Type":"ContainerStarted","Data":"65e522cedb817a208dc9b5ff4fb5219a5d041c1c534ed827d47015d6d2eac208"} Dec 01 18:02:04 crc kubenswrapper[4868]: I1201 18:02:04.809923 4868 generic.go:334] "Generic (PLEG): container finished" podID="0ba0b675-4358-4b02-81f1-d6223ad87229" containerID="65e522cedb817a208dc9b5ff4fb5219a5d041c1c534ed827d47015d6d2eac208" exitCode=0 Dec 01 18:02:04 crc kubenswrapper[4868]: I1201 18:02:04.810560 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2g2b" event={"ID":"0ba0b675-4358-4b02-81f1-d6223ad87229","Type":"ContainerDied","Data":"65e522cedb817a208dc9b5ff4fb5219a5d041c1c534ed827d47015d6d2eac208"} Dec 01 18:02:05 crc kubenswrapper[4868]: I1201 18:02:05.824100 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2g2b" event={"ID":"0ba0b675-4358-4b02-81f1-d6223ad87229","Type":"ContainerStarted","Data":"4d9600d0a6d5d538b10065620a4a59eed275c666727d88c341a811e401a8bb1d"} Dec 01 18:02:05 crc kubenswrapper[4868]: I1201 18:02:05.848802 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d2g2b" podStartSLOduration=2.066299951 podStartE2EDuration="7.848781534s" podCreationTimestamp="2025-12-01 18:01:58 +0000 UTC" firstStartedPulling="2025-12-01 18:01:59.768710563 +0000 UTC m=+2192.139820974" lastFinishedPulling="2025-12-01 18:02:05.551192156 +0000 UTC m=+2197.922302557" observedRunningTime="2025-12-01 18:02:05.844959571 +0000 UTC m=+2198.216069982" watchObservedRunningTime="2025-12-01 18:02:05.848781534 +0000 UTC m=+2198.219891945" Dec 01 18:02:08 crc kubenswrapper[4868]: I1201 18:02:08.523863 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:02:08 crc kubenswrapper[4868]: I1201 18:02:08.524198 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:02:08 crc kubenswrapper[4868]: I1201 18:02:08.682575 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fpxc7" Dec 01 18:02:08 crc kubenswrapper[4868]: I1201 18:02:08.729853 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fpxc7" Dec 01 18:02:08 crc kubenswrapper[4868]: I1201 18:02:08.786493 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fpxc7"] Dec 01 18:02:08 crc kubenswrapper[4868]: I1201 18:02:08.918402 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2b5vb"] Dec 01 18:02:08 crc kubenswrapper[4868]: I1201 18:02:08.918963 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2b5vb" podUID="05d6020b-0574-40f5-b421-fe6b21712fd2" containerName="registry-server" containerID="cri-o://721add12e4b66e4d21eb3e6db6f32a0316deb250a95a47c117ef73d74077d7c9" gracePeriod=2 Dec 01 18:02:09 crc kubenswrapper[4868]: E1201 18:02:09.106592 4868 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05d6020b_0574_40f5_b421_fe6b21712fd2.slice/crio-conmon-721add12e4b66e4d21eb3e6db6f32a0316deb250a95a47c117ef73d74077d7c9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05d6020b_0574_40f5_b421_fe6b21712fd2.slice/crio-721add12e4b66e4d21eb3e6db6f32a0316deb250a95a47c117ef73d74077d7c9.scope\": RecentStats: unable to find data in memory cache]" Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.448212 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2b5vb" Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.571499 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d2g2b" podUID="0ba0b675-4358-4b02-81f1-d6223ad87229" containerName="registry-server" probeResult="failure" output=< Dec 01 18:02:09 crc kubenswrapper[4868]: timeout: failed to connect service ":50051" within 1s Dec 01 18:02:09 crc kubenswrapper[4868]: > Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.585459 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05d6020b-0574-40f5-b421-fe6b21712fd2-catalog-content\") pod \"05d6020b-0574-40f5-b421-fe6b21712fd2\" (UID: \"05d6020b-0574-40f5-b421-fe6b21712fd2\") " Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.585993 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrjc2\" (UniqueName: \"kubernetes.io/projected/05d6020b-0574-40f5-b421-fe6b21712fd2-kube-api-access-hrjc2\") pod \"05d6020b-0574-40f5-b421-fe6b21712fd2\" (UID: \"05d6020b-0574-40f5-b421-fe6b21712fd2\") " Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.586892 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05d6020b-0574-40f5-b421-fe6b21712fd2-utilities\") pod \"05d6020b-0574-40f5-b421-fe6b21712fd2\" (UID: \"05d6020b-0574-40f5-b421-fe6b21712fd2\") " Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.588085 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05d6020b-0574-40f5-b421-fe6b21712fd2-utilities" (OuterVolumeSpecName: "utilities") pod "05d6020b-0574-40f5-b421-fe6b21712fd2" (UID: "05d6020b-0574-40f5-b421-fe6b21712fd2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.591794 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05d6020b-0574-40f5-b421-fe6b21712fd2-kube-api-access-hrjc2" (OuterVolumeSpecName: "kube-api-access-hrjc2") pod "05d6020b-0574-40f5-b421-fe6b21712fd2" (UID: "05d6020b-0574-40f5-b421-fe6b21712fd2"). InnerVolumeSpecName "kube-api-access-hrjc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.637035 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05d6020b-0574-40f5-b421-fe6b21712fd2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05d6020b-0574-40f5-b421-fe6b21712fd2" (UID: "05d6020b-0574-40f5-b421-fe6b21712fd2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.689301 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05d6020b-0574-40f5-b421-fe6b21712fd2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.689512 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrjc2\" (UniqueName: \"kubernetes.io/projected/05d6020b-0574-40f5-b421-fe6b21712fd2-kube-api-access-hrjc2\") on node \"crc\" DevicePath \"\"" Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.689576 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05d6020b-0574-40f5-b421-fe6b21712fd2-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.865526 4868 generic.go:334] "Generic (PLEG): container finished" podID="05d6020b-0574-40f5-b421-fe6b21712fd2" containerID="721add12e4b66e4d21eb3e6db6f32a0316deb250a95a47c117ef73d74077d7c9" exitCode=0 Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.865818 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2b5vb" Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.865839 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2b5vb" event={"ID":"05d6020b-0574-40f5-b421-fe6b21712fd2","Type":"ContainerDied","Data":"721add12e4b66e4d21eb3e6db6f32a0316deb250a95a47c117ef73d74077d7c9"} Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.866403 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2b5vb" event={"ID":"05d6020b-0574-40f5-b421-fe6b21712fd2","Type":"ContainerDied","Data":"4ab5b92cd7ad6e093f6dc67f3f54c0370389437bc68bd5adc53f3fc4c63c6906"} Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.866499 4868 scope.go:117] "RemoveContainer" containerID="721add12e4b66e4d21eb3e6db6f32a0316deb250a95a47c117ef73d74077d7c9" Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.900288 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2b5vb"] Dec 01 18:02:09 crc kubenswrapper[4868]: I1201 18:02:09.908780 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2b5vb"] Dec 01 18:02:10 crc kubenswrapper[4868]: I1201 18:02:10.075976 4868 scope.go:117] "RemoveContainer" containerID="c9e172b052a16138cfb8346293175e6ab32cbcc88b89d1be33a63fc0e8cbc65a" Dec 01 18:02:10 crc kubenswrapper[4868]: I1201 18:02:10.097005 4868 scope.go:117] "RemoveContainer" containerID="0c905699ffbd48633864afcc85b9b2c2159ade7d0493a9377771f6382d758587" Dec 01 18:02:10 crc kubenswrapper[4868]: I1201 18:02:10.141017 4868 scope.go:117] "RemoveContainer" containerID="721add12e4b66e4d21eb3e6db6f32a0316deb250a95a47c117ef73d74077d7c9" Dec 01 18:02:10 crc kubenswrapper[4868]: E1201 18:02:10.150446 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"721add12e4b66e4d21eb3e6db6f32a0316deb250a95a47c117ef73d74077d7c9\": container with ID starting with 721add12e4b66e4d21eb3e6db6f32a0316deb250a95a47c117ef73d74077d7c9 not found: ID does not exist" containerID="721add12e4b66e4d21eb3e6db6f32a0316deb250a95a47c117ef73d74077d7c9" Dec 01 18:02:10 crc kubenswrapper[4868]: I1201 18:02:10.150628 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"721add12e4b66e4d21eb3e6db6f32a0316deb250a95a47c117ef73d74077d7c9"} err="failed to get container status \"721add12e4b66e4d21eb3e6db6f32a0316deb250a95a47c117ef73d74077d7c9\": rpc error: code = NotFound desc = could not find container \"721add12e4b66e4d21eb3e6db6f32a0316deb250a95a47c117ef73d74077d7c9\": container with ID starting with 721add12e4b66e4d21eb3e6db6f32a0316deb250a95a47c117ef73d74077d7c9 not found: ID does not exist" Dec 01 18:02:10 crc kubenswrapper[4868]: I1201 18:02:10.150728 4868 scope.go:117] "RemoveContainer" containerID="c9e172b052a16138cfb8346293175e6ab32cbcc88b89d1be33a63fc0e8cbc65a" Dec 01 18:02:10 crc kubenswrapper[4868]: E1201 18:02:10.151245 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9e172b052a16138cfb8346293175e6ab32cbcc88b89d1be33a63fc0e8cbc65a\": container with ID starting with c9e172b052a16138cfb8346293175e6ab32cbcc88b89d1be33a63fc0e8cbc65a not found: ID does not exist" containerID="c9e172b052a16138cfb8346293175e6ab32cbcc88b89d1be33a63fc0e8cbc65a" Dec 01 18:02:10 crc kubenswrapper[4868]: I1201 18:02:10.151277 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9e172b052a16138cfb8346293175e6ab32cbcc88b89d1be33a63fc0e8cbc65a"} err="failed to get container status \"c9e172b052a16138cfb8346293175e6ab32cbcc88b89d1be33a63fc0e8cbc65a\": rpc error: code = NotFound desc = could not find container \"c9e172b052a16138cfb8346293175e6ab32cbcc88b89d1be33a63fc0e8cbc65a\": container with ID starting with c9e172b052a16138cfb8346293175e6ab32cbcc88b89d1be33a63fc0e8cbc65a not found: ID does not exist" Dec 01 18:02:10 crc kubenswrapper[4868]: I1201 18:02:10.151295 4868 scope.go:117] "RemoveContainer" containerID="0c905699ffbd48633864afcc85b9b2c2159ade7d0493a9377771f6382d758587" Dec 01 18:02:10 crc kubenswrapper[4868]: E1201 18:02:10.151666 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c905699ffbd48633864afcc85b9b2c2159ade7d0493a9377771f6382d758587\": container with ID starting with 0c905699ffbd48633864afcc85b9b2c2159ade7d0493a9377771f6382d758587 not found: ID does not exist" containerID="0c905699ffbd48633864afcc85b9b2c2159ade7d0493a9377771f6382d758587" Dec 01 18:02:10 crc kubenswrapper[4868]: I1201 18:02:10.151765 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c905699ffbd48633864afcc85b9b2c2159ade7d0493a9377771f6382d758587"} err="failed to get container status \"0c905699ffbd48633864afcc85b9b2c2159ade7d0493a9377771f6382d758587\": rpc error: code = NotFound desc = could not find container \"0c905699ffbd48633864afcc85b9b2c2159ade7d0493a9377771f6382d758587\": container with ID starting with 0c905699ffbd48633864afcc85b9b2c2159ade7d0493a9377771f6382d758587 not found: ID does not exist" Dec 01 18:02:10 crc kubenswrapper[4868]: I1201 18:02:10.183311 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05d6020b-0574-40f5-b421-fe6b21712fd2" path="/var/lib/kubelet/pods/05d6020b-0574-40f5-b421-fe6b21712fd2/volumes" Dec 01 18:02:18 crc kubenswrapper[4868]: I1201 18:02:18.574750 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:02:18 crc kubenswrapper[4868]: I1201 18:02:18.629668 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:02:19 crc kubenswrapper[4868]: I1201 18:02:19.479758 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d2g2b"] Dec 01 18:02:19 crc kubenswrapper[4868]: I1201 18:02:19.957313 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d2g2b" podUID="0ba0b675-4358-4b02-81f1-d6223ad87229" containerName="registry-server" containerID="cri-o://4d9600d0a6d5d538b10065620a4a59eed275c666727d88c341a811e401a8bb1d" gracePeriod=2 Dec 01 18:02:20 crc kubenswrapper[4868]: I1201 18:02:20.410640 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:02:20 crc kubenswrapper[4868]: I1201 18:02:20.506100 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5gmg\" (UniqueName: \"kubernetes.io/projected/0ba0b675-4358-4b02-81f1-d6223ad87229-kube-api-access-w5gmg\") pod \"0ba0b675-4358-4b02-81f1-d6223ad87229\" (UID: \"0ba0b675-4358-4b02-81f1-d6223ad87229\") " Dec 01 18:02:20 crc kubenswrapper[4868]: I1201 18:02:20.506255 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ba0b675-4358-4b02-81f1-d6223ad87229-catalog-content\") pod \"0ba0b675-4358-4b02-81f1-d6223ad87229\" (UID: \"0ba0b675-4358-4b02-81f1-d6223ad87229\") " Dec 01 18:02:20 crc kubenswrapper[4868]: I1201 18:02:20.506372 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ba0b675-4358-4b02-81f1-d6223ad87229-utilities\") pod \"0ba0b675-4358-4b02-81f1-d6223ad87229\" (UID: \"0ba0b675-4358-4b02-81f1-d6223ad87229\") " Dec 01 18:02:20 crc kubenswrapper[4868]: I1201 18:02:20.507524 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ba0b675-4358-4b02-81f1-d6223ad87229-utilities" (OuterVolumeSpecName: "utilities") pod "0ba0b675-4358-4b02-81f1-d6223ad87229" (UID: "0ba0b675-4358-4b02-81f1-d6223ad87229"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:02:20 crc kubenswrapper[4868]: I1201 18:02:20.513562 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ba0b675-4358-4b02-81f1-d6223ad87229-kube-api-access-w5gmg" (OuterVolumeSpecName: "kube-api-access-w5gmg") pod "0ba0b675-4358-4b02-81f1-d6223ad87229" (UID: "0ba0b675-4358-4b02-81f1-d6223ad87229"). InnerVolumeSpecName "kube-api-access-w5gmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:02:20 crc kubenswrapper[4868]: I1201 18:02:20.609095 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5gmg\" (UniqueName: \"kubernetes.io/projected/0ba0b675-4358-4b02-81f1-d6223ad87229-kube-api-access-w5gmg\") on node \"crc\" DevicePath \"\"" Dec 01 18:02:20 crc kubenswrapper[4868]: I1201 18:02:20.609171 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ba0b675-4358-4b02-81f1-d6223ad87229-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 18:02:20 crc kubenswrapper[4868]: I1201 18:02:20.640580 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ba0b675-4358-4b02-81f1-d6223ad87229-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ba0b675-4358-4b02-81f1-d6223ad87229" (UID: "0ba0b675-4358-4b02-81f1-d6223ad87229"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:02:20 crc kubenswrapper[4868]: I1201 18:02:20.710804 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ba0b675-4358-4b02-81f1-d6223ad87229-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 18:02:20 crc kubenswrapper[4868]: I1201 18:02:20.967440 4868 generic.go:334] "Generic (PLEG): container finished" podID="0ba0b675-4358-4b02-81f1-d6223ad87229" containerID="4d9600d0a6d5d538b10065620a4a59eed275c666727d88c341a811e401a8bb1d" exitCode=0 Dec 01 18:02:20 crc kubenswrapper[4868]: I1201 18:02:20.967493 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2g2b" event={"ID":"0ba0b675-4358-4b02-81f1-d6223ad87229","Type":"ContainerDied","Data":"4d9600d0a6d5d538b10065620a4a59eed275c666727d88c341a811e401a8bb1d"} Dec 01 18:02:20 crc kubenswrapper[4868]: I1201 18:02:20.967518 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2g2b" Dec 01 18:02:20 crc kubenswrapper[4868]: I1201 18:02:20.967544 4868 scope.go:117] "RemoveContainer" containerID="4d9600d0a6d5d538b10065620a4a59eed275c666727d88c341a811e401a8bb1d" Dec 01 18:02:20 crc kubenswrapper[4868]: I1201 18:02:20.967527 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2g2b" event={"ID":"0ba0b675-4358-4b02-81f1-d6223ad87229","Type":"ContainerDied","Data":"680a0e3d439d6fae9bb51a069da57e3510fba6e296b4800a68873f6ca21c60a3"} Dec 01 18:02:21 crc kubenswrapper[4868]: I1201 18:02:21.011120 4868 scope.go:117] "RemoveContainer" containerID="65e522cedb817a208dc9b5ff4fb5219a5d041c1c534ed827d47015d6d2eac208" Dec 01 18:02:21 crc kubenswrapper[4868]: I1201 18:02:21.051029 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d2g2b"] Dec 01 18:02:21 crc kubenswrapper[4868]: I1201 18:02:21.075324 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d2g2b"] Dec 01 18:02:21 crc kubenswrapper[4868]: I1201 18:02:21.079622 4868 scope.go:117] "RemoveContainer" containerID="8d05ee04fb47291b0a8cce8ea89cb6a4cdc53838eaac06b96d2bc1d57ebffd17" Dec 01 18:02:21 crc kubenswrapper[4868]: I1201 18:02:21.155340 4868 scope.go:117] "RemoveContainer" containerID="4d9600d0a6d5d538b10065620a4a59eed275c666727d88c341a811e401a8bb1d" Dec 01 18:02:21 crc kubenswrapper[4868]: E1201 18:02:21.158049 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d9600d0a6d5d538b10065620a4a59eed275c666727d88c341a811e401a8bb1d\": container with ID starting with 4d9600d0a6d5d538b10065620a4a59eed275c666727d88c341a811e401a8bb1d not found: ID does not exist" containerID="4d9600d0a6d5d538b10065620a4a59eed275c666727d88c341a811e401a8bb1d" Dec 01 18:02:21 crc kubenswrapper[4868]: I1201 18:02:21.158082 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d9600d0a6d5d538b10065620a4a59eed275c666727d88c341a811e401a8bb1d"} err="failed to get container status \"4d9600d0a6d5d538b10065620a4a59eed275c666727d88c341a811e401a8bb1d\": rpc error: code = NotFound desc = could not find container \"4d9600d0a6d5d538b10065620a4a59eed275c666727d88c341a811e401a8bb1d\": container with ID starting with 4d9600d0a6d5d538b10065620a4a59eed275c666727d88c341a811e401a8bb1d not found: ID does not exist" Dec 01 18:02:21 crc kubenswrapper[4868]: I1201 18:02:21.158123 4868 scope.go:117] "RemoveContainer" containerID="65e522cedb817a208dc9b5ff4fb5219a5d041c1c534ed827d47015d6d2eac208" Dec 01 18:02:21 crc kubenswrapper[4868]: E1201 18:02:21.158520 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65e522cedb817a208dc9b5ff4fb5219a5d041c1c534ed827d47015d6d2eac208\": container with ID starting with 65e522cedb817a208dc9b5ff4fb5219a5d041c1c534ed827d47015d6d2eac208 not found: ID does not exist" containerID="65e522cedb817a208dc9b5ff4fb5219a5d041c1c534ed827d47015d6d2eac208" Dec 01 18:02:21 crc kubenswrapper[4868]: I1201 18:02:21.158551 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65e522cedb817a208dc9b5ff4fb5219a5d041c1c534ed827d47015d6d2eac208"} err="failed to get container status \"65e522cedb817a208dc9b5ff4fb5219a5d041c1c534ed827d47015d6d2eac208\": rpc error: code = NotFound desc = could not find container \"65e522cedb817a208dc9b5ff4fb5219a5d041c1c534ed827d47015d6d2eac208\": container with ID starting with 65e522cedb817a208dc9b5ff4fb5219a5d041c1c534ed827d47015d6d2eac208 not found: ID does not exist" Dec 01 18:02:21 crc kubenswrapper[4868]: I1201 18:02:21.158569 4868 scope.go:117] "RemoveContainer" containerID="8d05ee04fb47291b0a8cce8ea89cb6a4cdc53838eaac06b96d2bc1d57ebffd17" Dec 01 18:02:21 crc kubenswrapper[4868]: E1201 18:02:21.158794 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d05ee04fb47291b0a8cce8ea89cb6a4cdc53838eaac06b96d2bc1d57ebffd17\": container with ID starting with 8d05ee04fb47291b0a8cce8ea89cb6a4cdc53838eaac06b96d2bc1d57ebffd17 not found: ID does not exist" containerID="8d05ee04fb47291b0a8cce8ea89cb6a4cdc53838eaac06b96d2bc1d57ebffd17" Dec 01 18:02:21 crc kubenswrapper[4868]: I1201 18:02:21.158816 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d05ee04fb47291b0a8cce8ea89cb6a4cdc53838eaac06b96d2bc1d57ebffd17"} err="failed to get container status \"8d05ee04fb47291b0a8cce8ea89cb6a4cdc53838eaac06b96d2bc1d57ebffd17\": rpc error: code = NotFound desc = could not find container \"8d05ee04fb47291b0a8cce8ea89cb6a4cdc53838eaac06b96d2bc1d57ebffd17\": container with ID starting with 8d05ee04fb47291b0a8cce8ea89cb6a4cdc53838eaac06b96d2bc1d57ebffd17 not found: ID does not exist" Dec 01 18:02:22 crc kubenswrapper[4868]: I1201 18:02:22.184693 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ba0b675-4358-4b02-81f1-d6223ad87229" path="/var/lib/kubelet/pods/0ba0b675-4358-4b02-81f1-d6223ad87229/volumes" Dec 01 18:02:47 crc kubenswrapper[4868]: I1201 18:02:47.210734 4868 generic.go:334] "Generic (PLEG): container finished" podID="944aa8d9-9aa5-466a-82fd-9768a8b970cb" containerID="734414eeaf22665634fee0caf486def154a10fa23059952ab26d09eede4e809e" exitCode=0 Dec 01 18:02:47 crc kubenswrapper[4868]: I1201 18:02:47.210841 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" event={"ID":"944aa8d9-9aa5-466a-82fd-9768a8b970cb","Type":"ContainerDied","Data":"734414eeaf22665634fee0caf486def154a10fa23059952ab26d09eede4e809e"} Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.622581 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.768818 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkfrl\" (UniqueName: \"kubernetes.io/projected/944aa8d9-9aa5-466a-82fd-9768a8b970cb-kube-api-access-hkfrl\") pod \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.768892 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-ssh-key\") pod \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.769104 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-libvirt-combined-ca-bundle\") pod \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.769130 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-inventory\") pod \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.769179 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-libvirt-secret-0\") pod \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\" (UID: \"944aa8d9-9aa5-466a-82fd-9768a8b970cb\") " Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.787146 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "944aa8d9-9aa5-466a-82fd-9768a8b970cb" (UID: "944aa8d9-9aa5-466a-82fd-9768a8b970cb"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.787159 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/944aa8d9-9aa5-466a-82fd-9768a8b970cb-kube-api-access-hkfrl" (OuterVolumeSpecName: "kube-api-access-hkfrl") pod "944aa8d9-9aa5-466a-82fd-9768a8b970cb" (UID: "944aa8d9-9aa5-466a-82fd-9768a8b970cb"). InnerVolumeSpecName "kube-api-access-hkfrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.796156 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "944aa8d9-9aa5-466a-82fd-9768a8b970cb" (UID: "944aa8d9-9aa5-466a-82fd-9768a8b970cb"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.796240 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "944aa8d9-9aa5-466a-82fd-9768a8b970cb" (UID: "944aa8d9-9aa5-466a-82fd-9768a8b970cb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.798312 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-inventory" (OuterVolumeSpecName: "inventory") pod "944aa8d9-9aa5-466a-82fd-9768a8b970cb" (UID: "944aa8d9-9aa5-466a-82fd-9768a8b970cb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.871531 4868 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.871708 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.871787 4868 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.871912 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkfrl\" (UniqueName: \"kubernetes.io/projected/944aa8d9-9aa5-466a-82fd-9768a8b970cb-kube-api-access-hkfrl\") on node \"crc\" DevicePath \"\"" Dec 01 18:02:48 crc kubenswrapper[4868]: I1201 18:02:48.872015 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/944aa8d9-9aa5-466a-82fd-9768a8b970cb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.237264 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" event={"ID":"944aa8d9-9aa5-466a-82fd-9768a8b970cb","Type":"ContainerDied","Data":"b2351a5ee6f05528c2017db48e8ec349058c583f5d1a94993d57f05827a36256"} Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.237305 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2351a5ee6f05528c2017db48e8ec349058c583f5d1a94993d57f05827a36256" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.237323 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.340880 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm"] Dec 01 18:02:49 crc kubenswrapper[4868]: E1201 18:02:49.341322 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b880bd-ac45-4fee-9a95-9fbd38285f18" containerName="extract-utilities" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.341341 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b880bd-ac45-4fee-9a95-9fbd38285f18" containerName="extract-utilities" Dec 01 18:02:49 crc kubenswrapper[4868]: E1201 18:02:49.341363 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ba0b675-4358-4b02-81f1-d6223ad87229" containerName="extract-content" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.341370 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ba0b675-4358-4b02-81f1-d6223ad87229" containerName="extract-content" Dec 01 18:02:49 crc kubenswrapper[4868]: E1201 18:02:49.341382 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05d6020b-0574-40f5-b421-fe6b21712fd2" containerName="registry-server" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.341388 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="05d6020b-0574-40f5-b421-fe6b21712fd2" containerName="registry-server" Dec 01 18:02:49 crc kubenswrapper[4868]: E1201 18:02:49.341400 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="944aa8d9-9aa5-466a-82fd-9768a8b970cb" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.341407 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="944aa8d9-9aa5-466a-82fd-9768a8b970cb" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 18:02:49 crc kubenswrapper[4868]: E1201 18:02:49.341423 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05d6020b-0574-40f5-b421-fe6b21712fd2" containerName="extract-content" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.341430 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="05d6020b-0574-40f5-b421-fe6b21712fd2" containerName="extract-content" Dec 01 18:02:49 crc kubenswrapper[4868]: E1201 18:02:49.341448 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b880bd-ac45-4fee-9a95-9fbd38285f18" containerName="registry-server" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.341454 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b880bd-ac45-4fee-9a95-9fbd38285f18" containerName="registry-server" Dec 01 18:02:49 crc kubenswrapper[4868]: E1201 18:02:49.341465 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ba0b675-4358-4b02-81f1-d6223ad87229" containerName="registry-server" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.341470 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ba0b675-4358-4b02-81f1-d6223ad87229" containerName="registry-server" Dec 01 18:02:49 crc kubenswrapper[4868]: E1201 18:02:49.341483 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b880bd-ac45-4fee-9a95-9fbd38285f18" containerName="extract-content" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.341488 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b880bd-ac45-4fee-9a95-9fbd38285f18" containerName="extract-content" Dec 01 18:02:49 crc kubenswrapper[4868]: E1201 18:02:49.341507 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ba0b675-4358-4b02-81f1-d6223ad87229" containerName="extract-utilities" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.341512 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ba0b675-4358-4b02-81f1-d6223ad87229" containerName="extract-utilities" Dec 01 18:02:49 crc kubenswrapper[4868]: E1201 18:02:49.341519 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05d6020b-0574-40f5-b421-fe6b21712fd2" containerName="extract-utilities" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.341524 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="05d6020b-0574-40f5-b421-fe6b21712fd2" containerName="extract-utilities" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.341692 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="944aa8d9-9aa5-466a-82fd-9768a8b970cb" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.341712 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="05d6020b-0574-40f5-b421-fe6b21712fd2" containerName="registry-server" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.341724 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4b880bd-ac45-4fee-9a95-9fbd38285f18" containerName="registry-server" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.341746 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ba0b675-4358-4b02-81f1-d6223ad87229" containerName="registry-server" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.342384 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.344520 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.344555 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.344534 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.344535 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.345172 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.345354 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.348390 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.356841 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm"] Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.383383 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.383535 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.383576 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.383605 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.383654 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.383683 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.383718 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.383758 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.383782 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xwm7\" (UniqueName: \"kubernetes.io/projected/0739d3f5-f361-46f0-872d-23f9d29b7d46-kube-api-access-2xwm7\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.485977 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.486325 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.486350 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.486369 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.486406 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.486427 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.486458 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.486486 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.486508 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xwm7\" (UniqueName: \"kubernetes.io/projected/0739d3f5-f361-46f0-872d-23f9d29b7d46-kube-api-access-2xwm7\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.487259 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.490401 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.490485 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.491173 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.491535 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.492025 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.492782 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.496746 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.502424 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xwm7\" (UniqueName: \"kubernetes.io/projected/0739d3f5-f361-46f0-872d-23f9d29b7d46-kube-api-access-2xwm7\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6jdm\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:49 crc kubenswrapper[4868]: I1201 18:02:49.658860 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:02:50 crc kubenswrapper[4868]: I1201 18:02:50.042755 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm"] Dec 01 18:02:50 crc kubenswrapper[4868]: I1201 18:02:50.246756 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" event={"ID":"0739d3f5-f361-46f0-872d-23f9d29b7d46","Type":"ContainerStarted","Data":"1b3d3a4c623c9fdac5ed8d450394bdef19696f2efd1c981b1b2c7f25a336df94"} Dec 01 18:02:51 crc kubenswrapper[4868]: I1201 18:02:51.258310 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" event={"ID":"0739d3f5-f361-46f0-872d-23f9d29b7d46","Type":"ContainerStarted","Data":"1bf961625c8d79c02172cd0c5587dd27448abba703125915b7458966ce872413"} Dec 01 18:02:51 crc kubenswrapper[4868]: I1201 18:02:51.276395 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" podStartSLOduration=1.755049964 podStartE2EDuration="2.276291902s" podCreationTimestamp="2025-12-01 18:02:49 +0000 UTC" firstStartedPulling="2025-12-01 18:02:50.047146982 +0000 UTC m=+2242.418257393" lastFinishedPulling="2025-12-01 18:02:50.56838892 +0000 UTC m=+2242.939499331" observedRunningTime="2025-12-01 18:02:51.273858805 +0000 UTC m=+2243.644969226" watchObservedRunningTime="2025-12-01 18:02:51.276291902 +0000 UTC m=+2243.647402333" Dec 01 18:04:25 crc kubenswrapper[4868]: I1201 18:04:25.904885 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:04:25 crc kubenswrapper[4868]: I1201 18:04:25.905380 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:04:55 crc kubenswrapper[4868]: I1201 18:04:55.904806 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:04:55 crc kubenswrapper[4868]: I1201 18:04:55.905502 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:05:25 crc kubenswrapper[4868]: I1201 18:05:25.904542 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:05:25 crc kubenswrapper[4868]: I1201 18:05:25.905491 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:05:25 crc kubenswrapper[4868]: I1201 18:05:25.905571 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 18:05:25 crc kubenswrapper[4868]: I1201 18:05:25.906843 4868 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892"} pod="openshift-machine-config-operator/machine-config-daemon-njgpn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 18:05:25 crc kubenswrapper[4868]: I1201 18:05:25.906923 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" containerID="cri-o://78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" gracePeriod=600 Dec 01 18:05:26 crc kubenswrapper[4868]: E1201 18:05:26.039392 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:05:26 crc kubenswrapper[4868]: I1201 18:05:26.826735 4868 generic.go:334] "Generic (PLEG): container finished" podID="ec4e08bf-df56-4490-8e79-a4175c727405" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" exitCode=0 Dec 01 18:05:26 crc kubenswrapper[4868]: I1201 18:05:26.827129 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerDied","Data":"78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892"} Dec 01 18:05:26 crc kubenswrapper[4868]: I1201 18:05:26.827174 4868 scope.go:117] "RemoveContainer" containerID="95b0a3a129d931e7326436a081957bc1421283eca7e6144750f7284aee7a3835" Dec 01 18:05:26 crc kubenswrapper[4868]: I1201 18:05:26.827981 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:05:26 crc kubenswrapper[4868]: E1201 18:05:26.828300 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:05:27 crc kubenswrapper[4868]: I1201 18:05:27.840241 4868 generic.go:334] "Generic (PLEG): container finished" podID="0739d3f5-f361-46f0-872d-23f9d29b7d46" containerID="1bf961625c8d79c02172cd0c5587dd27448abba703125915b7458966ce872413" exitCode=0 Dec 01 18:05:27 crc kubenswrapper[4868]: I1201 18:05:27.840283 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" event={"ID":"0739d3f5-f361-46f0-872d-23f9d29b7d46","Type":"ContainerDied","Data":"1bf961625c8d79c02172cd0c5587dd27448abba703125915b7458966ce872413"} Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.272212 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.439813 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-cell1-compute-config-1\") pod \"0739d3f5-f361-46f0-872d-23f9d29b7d46\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.439880 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-combined-ca-bundle\") pod \"0739d3f5-f361-46f0-872d-23f9d29b7d46\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.439922 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-cell1-compute-config-0\") pod \"0739d3f5-f361-46f0-872d-23f9d29b7d46\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.439995 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-migration-ssh-key-1\") pod \"0739d3f5-f361-46f0-872d-23f9d29b7d46\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.440089 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-ssh-key\") pod \"0739d3f5-f361-46f0-872d-23f9d29b7d46\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.440123 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-migration-ssh-key-0\") pod \"0739d3f5-f361-46f0-872d-23f9d29b7d46\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.440144 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xwm7\" (UniqueName: \"kubernetes.io/projected/0739d3f5-f361-46f0-872d-23f9d29b7d46-kube-api-access-2xwm7\") pod \"0739d3f5-f361-46f0-872d-23f9d29b7d46\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.440192 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-extra-config-0\") pod \"0739d3f5-f361-46f0-872d-23f9d29b7d46\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.440241 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-inventory\") pod \"0739d3f5-f361-46f0-872d-23f9d29b7d46\" (UID: \"0739d3f5-f361-46f0-872d-23f9d29b7d46\") " Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.446566 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0739d3f5-f361-46f0-872d-23f9d29b7d46-kube-api-access-2xwm7" (OuterVolumeSpecName: "kube-api-access-2xwm7") pod "0739d3f5-f361-46f0-872d-23f9d29b7d46" (UID: "0739d3f5-f361-46f0-872d-23f9d29b7d46"). InnerVolumeSpecName "kube-api-access-2xwm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.452336 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "0739d3f5-f361-46f0-872d-23f9d29b7d46" (UID: "0739d3f5-f361-46f0-872d-23f9d29b7d46"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.471048 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "0739d3f5-f361-46f0-872d-23f9d29b7d46" (UID: "0739d3f5-f361-46f0-872d-23f9d29b7d46"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.473217 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "0739d3f5-f361-46f0-872d-23f9d29b7d46" (UID: "0739d3f5-f361-46f0-872d-23f9d29b7d46"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.473680 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0739d3f5-f361-46f0-872d-23f9d29b7d46" (UID: "0739d3f5-f361-46f0-872d-23f9d29b7d46"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.478931 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "0739d3f5-f361-46f0-872d-23f9d29b7d46" (UID: "0739d3f5-f361-46f0-872d-23f9d29b7d46"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.480291 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "0739d3f5-f361-46f0-872d-23f9d29b7d46" (UID: "0739d3f5-f361-46f0-872d-23f9d29b7d46"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.485394 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "0739d3f5-f361-46f0-872d-23f9d29b7d46" (UID: "0739d3f5-f361-46f0-872d-23f9d29b7d46"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.492283 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-inventory" (OuterVolumeSpecName: "inventory") pod "0739d3f5-f361-46f0-872d-23f9d29b7d46" (UID: "0739d3f5-f361-46f0-872d-23f9d29b7d46"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.542750 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.542909 4868 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.543002 4868 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.543090 4868 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.543150 4868 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.543203 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.543265 4868 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.543320 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xwm7\" (UniqueName: \"kubernetes.io/projected/0739d3f5-f361-46f0-872d-23f9d29b7d46-kube-api-access-2xwm7\") on node \"crc\" DevicePath \"\"" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.543401 4868 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/0739d3f5-f361-46f0-872d-23f9d29b7d46-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.857917 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" event={"ID":"0739d3f5-f361-46f0-872d-23f9d29b7d46","Type":"ContainerDied","Data":"1b3d3a4c623c9fdac5ed8d450394bdef19696f2efd1c981b1b2c7f25a336df94"} Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.857979 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b3d3a4c623c9fdac5ed8d450394bdef19696f2efd1c981b1b2c7f25a336df94" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.858023 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6jdm" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.973777 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq"] Dec 01 18:05:29 crc kubenswrapper[4868]: E1201 18:05:29.974596 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0739d3f5-f361-46f0-872d-23f9d29b7d46" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.974619 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="0739d3f5-f361-46f0-872d-23f9d29b7d46" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.974906 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="0739d3f5-f361-46f0-872d-23f9d29b7d46" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.975737 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.978131 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.980014 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.980297 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.980463 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ht542" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.980720 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 18:05:29 crc kubenswrapper[4868]: I1201 18:05:29.982847 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq"] Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.050992 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.051079 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.051274 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpzck\" (UniqueName: \"kubernetes.io/projected/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-kube-api-access-tpzck\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.051443 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.051562 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.051591 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.051643 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.152860 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.152998 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.153071 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.153113 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpzck\" (UniqueName: \"kubernetes.io/projected/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-kube-api-access-tpzck\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.153181 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.153234 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.153258 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.157921 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.157932 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.157973 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.158190 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.158319 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.163484 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.173050 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpzck\" (UniqueName: \"kubernetes.io/projected/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-kube-api-access-tpzck\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.298247 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:05:30 crc kubenswrapper[4868]: I1201 18:05:30.856174 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq"] Dec 01 18:05:30 crc kubenswrapper[4868]: W1201 18:05:30.858177 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64998ba8_012c_4ffc_a7b3_dbbbb41940c0.slice/crio-b1644b0badd3d6ff44ba7803f7b09d7dee52dcd72cf099e14cbc310d568d85e8 WatchSource:0}: Error finding container b1644b0badd3d6ff44ba7803f7b09d7dee52dcd72cf099e14cbc310d568d85e8: Status 404 returned error can't find the container with id b1644b0badd3d6ff44ba7803f7b09d7dee52dcd72cf099e14cbc310d568d85e8 Dec 01 18:05:31 crc kubenswrapper[4868]: I1201 18:05:31.875862 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" event={"ID":"64998ba8-012c-4ffc-a7b3-dbbbb41940c0","Type":"ContainerStarted","Data":"e39bd285e71a67b9ec3d4f29fceb1a9fd27a424cc5577845e63bcba0c60ff4fd"} Dec 01 18:05:31 crc kubenswrapper[4868]: I1201 18:05:31.877126 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" event={"ID":"64998ba8-012c-4ffc-a7b3-dbbbb41940c0","Type":"ContainerStarted","Data":"b1644b0badd3d6ff44ba7803f7b09d7dee52dcd72cf099e14cbc310d568d85e8"} Dec 01 18:05:31 crc kubenswrapper[4868]: I1201 18:05:31.895551 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" podStartSLOduration=2.16405424 podStartE2EDuration="2.895525479s" podCreationTimestamp="2025-12-01 18:05:29 +0000 UTC" firstStartedPulling="2025-12-01 18:05:30.86036788 +0000 UTC m=+2403.231478291" lastFinishedPulling="2025-12-01 18:05:31.591839119 +0000 UTC m=+2403.962949530" observedRunningTime="2025-12-01 18:05:31.893642918 +0000 UTC m=+2404.264753329" watchObservedRunningTime="2025-12-01 18:05:31.895525479 +0000 UTC m=+2404.266635900" Dec 01 18:05:39 crc kubenswrapper[4868]: I1201 18:05:39.172936 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:05:39 crc kubenswrapper[4868]: E1201 18:05:39.173718 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:05:50 crc kubenswrapper[4868]: I1201 18:05:50.172408 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:05:50 crc kubenswrapper[4868]: E1201 18:05:50.173221 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:06:05 crc kubenswrapper[4868]: I1201 18:06:05.172262 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:06:05 crc kubenswrapper[4868]: E1201 18:06:05.172959 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:06:16 crc kubenswrapper[4868]: I1201 18:06:16.172319 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:06:16 crc kubenswrapper[4868]: E1201 18:06:16.173265 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:06:27 crc kubenswrapper[4868]: I1201 18:06:27.172269 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:06:27 crc kubenswrapper[4868]: E1201 18:06:27.173155 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:06:41 crc kubenswrapper[4868]: I1201 18:06:41.171861 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:06:41 crc kubenswrapper[4868]: E1201 18:06:41.172842 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:06:54 crc kubenswrapper[4868]: I1201 18:06:54.172487 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:06:54 crc kubenswrapper[4868]: E1201 18:06:54.173352 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:07:09 crc kubenswrapper[4868]: I1201 18:07:09.172466 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:07:09 crc kubenswrapper[4868]: E1201 18:07:09.173241 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:07:21 crc kubenswrapper[4868]: I1201 18:07:21.172372 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:07:21 crc kubenswrapper[4868]: E1201 18:07:21.173175 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:07:32 crc kubenswrapper[4868]: I1201 18:07:32.172156 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:07:32 crc kubenswrapper[4868]: E1201 18:07:32.172910 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:07:39 crc kubenswrapper[4868]: I1201 18:07:39.096317 4868 generic.go:334] "Generic (PLEG): container finished" podID="64998ba8-012c-4ffc-a7b3-dbbbb41940c0" containerID="e39bd285e71a67b9ec3d4f29fceb1a9fd27a424cc5577845e63bcba0c60ff4fd" exitCode=0 Dec 01 18:07:39 crc kubenswrapper[4868]: I1201 18:07:39.096396 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" event={"ID":"64998ba8-012c-4ffc-a7b3-dbbbb41940c0","Type":"ContainerDied","Data":"e39bd285e71a67b9ec3d4f29fceb1a9fd27a424cc5577845e63bcba0c60ff4fd"} Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.539182 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.716031 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ssh-key\") pod \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.716079 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpzck\" (UniqueName: \"kubernetes.io/projected/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-kube-api-access-tpzck\") pod \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.716133 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-inventory\") pod \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.716179 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-2\") pod \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.716217 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-telemetry-combined-ca-bundle\") pod \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.716303 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-1\") pod \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.716360 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-0\") pod \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\" (UID: \"64998ba8-012c-4ffc-a7b3-dbbbb41940c0\") " Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.721597 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "64998ba8-012c-4ffc-a7b3-dbbbb41940c0" (UID: "64998ba8-012c-4ffc-a7b3-dbbbb41940c0"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.723348 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-kube-api-access-tpzck" (OuterVolumeSpecName: "kube-api-access-tpzck") pod "64998ba8-012c-4ffc-a7b3-dbbbb41940c0" (UID: "64998ba8-012c-4ffc-a7b3-dbbbb41940c0"). InnerVolumeSpecName "kube-api-access-tpzck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.742833 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "64998ba8-012c-4ffc-a7b3-dbbbb41940c0" (UID: "64998ba8-012c-4ffc-a7b3-dbbbb41940c0"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.743114 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "64998ba8-012c-4ffc-a7b3-dbbbb41940c0" (UID: "64998ba8-012c-4ffc-a7b3-dbbbb41940c0"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.743164 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "64998ba8-012c-4ffc-a7b3-dbbbb41940c0" (UID: "64998ba8-012c-4ffc-a7b3-dbbbb41940c0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.749887 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-inventory" (OuterVolumeSpecName: "inventory") pod "64998ba8-012c-4ffc-a7b3-dbbbb41940c0" (UID: "64998ba8-012c-4ffc-a7b3-dbbbb41940c0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.751101 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "64998ba8-012c-4ffc-a7b3-dbbbb41940c0" (UID: "64998ba8-012c-4ffc-a7b3-dbbbb41940c0"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.818341 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.818373 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpzck\" (UniqueName: \"kubernetes.io/projected/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-kube-api-access-tpzck\") on node \"crc\" DevicePath \"\"" Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.818388 4868 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.818400 4868 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.818411 4868 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.818452 4868 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 01 18:07:40 crc kubenswrapper[4868]: I1201 18:07:40.818466 4868 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/64998ba8-012c-4ffc-a7b3-dbbbb41940c0-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 01 18:07:41 crc kubenswrapper[4868]: I1201 18:07:41.114285 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" event={"ID":"64998ba8-012c-4ffc-a7b3-dbbbb41940c0","Type":"ContainerDied","Data":"b1644b0badd3d6ff44ba7803f7b09d7dee52dcd72cf099e14cbc310d568d85e8"} Dec 01 18:07:41 crc kubenswrapper[4868]: I1201 18:07:41.114321 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1644b0badd3d6ff44ba7803f7b09d7dee52dcd72cf099e14cbc310d568d85e8" Dec 01 18:07:41 crc kubenswrapper[4868]: I1201 18:07:41.114619 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq" Dec 01 18:07:43 crc kubenswrapper[4868]: I1201 18:07:43.172191 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:07:43 crc kubenswrapper[4868]: E1201 18:07:43.172728 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:07:57 crc kubenswrapper[4868]: I1201 18:07:57.171971 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:07:57 crc kubenswrapper[4868]: E1201 18:07:57.172773 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:08:09 crc kubenswrapper[4868]: I1201 18:08:09.171705 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:08:09 crc kubenswrapper[4868]: E1201 18:08:09.172797 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:08:20 crc kubenswrapper[4868]: I1201 18:08:20.172316 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:08:20 crc kubenswrapper[4868]: E1201 18:08:20.173324 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:08:31 crc kubenswrapper[4868]: I1201 18:08:31.172837 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:08:31 crc kubenswrapper[4868]: E1201 18:08:31.173889 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.545400 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 01 18:08:32 crc kubenswrapper[4868]: E1201 18:08:32.546096 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64998ba8-012c-4ffc-a7b3-dbbbb41940c0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.546112 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="64998ba8-012c-4ffc-a7b3-dbbbb41940c0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.546300 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="64998ba8-012c-4ffc-a7b3-dbbbb41940c0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.546907 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.549351 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.549387 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vbmbb" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.549358 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.549575 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.577057 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.640143 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8447d0d7-b9fd-4a48-bf54-e43272539aec-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.640423 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.640543 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.640670 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szcfg\" (UniqueName: \"kubernetes.io/projected/8447d0d7-b9fd-4a48-bf54-e43272539aec-kube-api-access-szcfg\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.640795 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8447d0d7-b9fd-4a48-bf54-e43272539aec-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.640917 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.641105 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8447d0d7-b9fd-4a48-bf54-e43272539aec-config-data\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.641194 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.641304 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8447d0d7-b9fd-4a48-bf54-e43272539aec-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.742989 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szcfg\" (UniqueName: \"kubernetes.io/projected/8447d0d7-b9fd-4a48-bf54-e43272539aec-kube-api-access-szcfg\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.743083 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8447d0d7-b9fd-4a48-bf54-e43272539aec-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.743125 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.743155 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8447d0d7-b9fd-4a48-bf54-e43272539aec-config-data\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.743180 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.743206 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8447d0d7-b9fd-4a48-bf54-e43272539aec-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.743312 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8447d0d7-b9fd-4a48-bf54-e43272539aec-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.743342 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.743401 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.744494 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8447d0d7-b9fd-4a48-bf54-e43272539aec-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.744646 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8447d0d7-b9fd-4a48-bf54-e43272539aec-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.744855 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.745744 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8447d0d7-b9fd-4a48-bf54-e43272539aec-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.745885 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8447d0d7-b9fd-4a48-bf54-e43272539aec-config-data\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.750585 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.751324 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.751738 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.768802 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szcfg\" (UniqueName: \"kubernetes.io/projected/8447d0d7-b9fd-4a48-bf54-e43272539aec-kube-api-access-szcfg\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.779315 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " pod="openstack/tempest-tests-tempest" Dec 01 18:08:32 crc kubenswrapper[4868]: I1201 18:08:32.875162 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 18:08:33 crc kubenswrapper[4868]: I1201 18:08:33.338322 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 01 18:08:33 crc kubenswrapper[4868]: I1201 18:08:33.348803 4868 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 18:08:33 crc kubenswrapper[4868]: I1201 18:08:33.607156 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8447d0d7-b9fd-4a48-bf54-e43272539aec","Type":"ContainerStarted","Data":"695d6b3cf3d3bb90fb02b3e897b8e6f0c33e3b33663d5cb8dfd78c1c871a7104"} Dec 01 18:08:45 crc kubenswrapper[4868]: I1201 18:08:45.172802 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:08:45 crc kubenswrapper[4868]: E1201 18:08:45.173924 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:08:58 crc kubenswrapper[4868]: I1201 18:08:58.186789 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:08:58 crc kubenswrapper[4868]: E1201 18:08:58.187674 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:09:00 crc kubenswrapper[4868]: E1201 18:09:00.673218 4868 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 01 18:09:00 crc kubenswrapper[4868]: E1201 18:09:00.673713 4868 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-szcfg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(8447d0d7-b9fd-4a48-bf54-e43272539aec): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 18:09:00 crc kubenswrapper[4868]: E1201 18:09:00.675499 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="8447d0d7-b9fd-4a48-bf54-e43272539aec" Dec 01 18:09:00 crc kubenswrapper[4868]: E1201 18:09:00.926370 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="8447d0d7-b9fd-4a48-bf54-e43272539aec" Dec 01 18:09:13 crc kubenswrapper[4868]: I1201 18:09:13.172413 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:09:13 crc kubenswrapper[4868]: E1201 18:09:13.173971 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:09:16 crc kubenswrapper[4868]: I1201 18:09:16.147792 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8447d0d7-b9fd-4a48-bf54-e43272539aec","Type":"ContainerStarted","Data":"ad95d425eea60265bc1bc6b17537f4fba2176f26007dd3eff5ef8b09335af438"} Dec 01 18:09:16 crc kubenswrapper[4868]: I1201 18:09:16.171206 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.890258477 podStartE2EDuration="45.17118773s" podCreationTimestamp="2025-12-01 18:08:31 +0000 UTC" firstStartedPulling="2025-12-01 18:08:33.348392278 +0000 UTC m=+2585.719502719" lastFinishedPulling="2025-12-01 18:09:14.629321521 +0000 UTC m=+2627.000431972" observedRunningTime="2025-12-01 18:09:16.169122624 +0000 UTC m=+2628.540233055" watchObservedRunningTime="2025-12-01 18:09:16.17118773 +0000 UTC m=+2628.542298151" Dec 01 18:09:26 crc kubenswrapper[4868]: I1201 18:09:26.172407 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:09:26 crc kubenswrapper[4868]: E1201 18:09:26.173382 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:09:41 crc kubenswrapper[4868]: I1201 18:09:41.171763 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:09:41 crc kubenswrapper[4868]: E1201 18:09:41.172508 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:09:54 crc kubenswrapper[4868]: I1201 18:09:54.172395 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:09:54 crc kubenswrapper[4868]: E1201 18:09:54.173276 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:10:09 crc kubenswrapper[4868]: I1201 18:10:09.172896 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:10:09 crc kubenswrapper[4868]: E1201 18:10:09.174175 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:10:21 crc kubenswrapper[4868]: I1201 18:10:21.172005 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:10:21 crc kubenswrapper[4868]: E1201 18:10:21.172725 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:10:34 crc kubenswrapper[4868]: I1201 18:10:34.173047 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:10:34 crc kubenswrapper[4868]: I1201 18:10:34.874398 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"a92c978799e731b7e9c55a869165d4a8aa38fccf54030879f3328db6186e34c0"} Dec 01 18:10:36 crc kubenswrapper[4868]: I1201 18:10:36.373574 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q2dkl"] Dec 01 18:10:36 crc kubenswrapper[4868]: I1201 18:10:36.380645 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:36 crc kubenswrapper[4868]: I1201 18:10:36.384518 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q2dkl"] Dec 01 18:10:36 crc kubenswrapper[4868]: I1201 18:10:36.467183 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cc579ef-54f6-41a8-83c9-27830c9602bc-catalog-content\") pod \"certified-operators-q2dkl\" (UID: \"3cc579ef-54f6-41a8-83c9-27830c9602bc\") " pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:36 crc kubenswrapper[4868]: I1201 18:10:36.467304 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjk5g\" (UniqueName: \"kubernetes.io/projected/3cc579ef-54f6-41a8-83c9-27830c9602bc-kube-api-access-fjk5g\") pod \"certified-operators-q2dkl\" (UID: \"3cc579ef-54f6-41a8-83c9-27830c9602bc\") " pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:36 crc kubenswrapper[4868]: I1201 18:10:36.467428 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cc579ef-54f6-41a8-83c9-27830c9602bc-utilities\") pod \"certified-operators-q2dkl\" (UID: \"3cc579ef-54f6-41a8-83c9-27830c9602bc\") " pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:36 crc kubenswrapper[4868]: I1201 18:10:36.569403 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cc579ef-54f6-41a8-83c9-27830c9602bc-utilities\") pod \"certified-operators-q2dkl\" (UID: \"3cc579ef-54f6-41a8-83c9-27830c9602bc\") " pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:36 crc kubenswrapper[4868]: I1201 18:10:36.569501 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cc579ef-54f6-41a8-83c9-27830c9602bc-catalog-content\") pod \"certified-operators-q2dkl\" (UID: \"3cc579ef-54f6-41a8-83c9-27830c9602bc\") " pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:36 crc kubenswrapper[4868]: I1201 18:10:36.569552 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjk5g\" (UniqueName: \"kubernetes.io/projected/3cc579ef-54f6-41a8-83c9-27830c9602bc-kube-api-access-fjk5g\") pod \"certified-operators-q2dkl\" (UID: \"3cc579ef-54f6-41a8-83c9-27830c9602bc\") " pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:36 crc kubenswrapper[4868]: I1201 18:10:36.570142 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cc579ef-54f6-41a8-83c9-27830c9602bc-utilities\") pod \"certified-operators-q2dkl\" (UID: \"3cc579ef-54f6-41a8-83c9-27830c9602bc\") " pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:36 crc kubenswrapper[4868]: I1201 18:10:36.570355 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cc579ef-54f6-41a8-83c9-27830c9602bc-catalog-content\") pod \"certified-operators-q2dkl\" (UID: \"3cc579ef-54f6-41a8-83c9-27830c9602bc\") " pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:36 crc kubenswrapper[4868]: I1201 18:10:36.589671 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjk5g\" (UniqueName: \"kubernetes.io/projected/3cc579ef-54f6-41a8-83c9-27830c9602bc-kube-api-access-fjk5g\") pod \"certified-operators-q2dkl\" (UID: \"3cc579ef-54f6-41a8-83c9-27830c9602bc\") " pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:36 crc kubenswrapper[4868]: I1201 18:10:36.722605 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:37 crc kubenswrapper[4868]: I1201 18:10:37.269053 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q2dkl"] Dec 01 18:10:37 crc kubenswrapper[4868]: I1201 18:10:37.922262 4868 generic.go:334] "Generic (PLEG): container finished" podID="3cc579ef-54f6-41a8-83c9-27830c9602bc" containerID="e0ea49e65ca32fa4561b8c826c3ea016fcb6bb9914cd07b0cd8f0413da885d2a" exitCode=0 Dec 01 18:10:37 crc kubenswrapper[4868]: I1201 18:10:37.922326 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2dkl" event={"ID":"3cc579ef-54f6-41a8-83c9-27830c9602bc","Type":"ContainerDied","Data":"e0ea49e65ca32fa4561b8c826c3ea016fcb6bb9914cd07b0cd8f0413da885d2a"} Dec 01 18:10:37 crc kubenswrapper[4868]: I1201 18:10:37.922653 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2dkl" event={"ID":"3cc579ef-54f6-41a8-83c9-27830c9602bc","Type":"ContainerStarted","Data":"9029f56fbd961b5980625c9c321849e36cc4620beb23af7def1f2eab2daa952f"} Dec 01 18:10:38 crc kubenswrapper[4868]: I1201 18:10:38.935068 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2dkl" event={"ID":"3cc579ef-54f6-41a8-83c9-27830c9602bc","Type":"ContainerStarted","Data":"965b279256caf6c73f7fc9e2cb17f071e4fcf4fe66d7f0b24beda345c4562d91"} Dec 01 18:10:39 crc kubenswrapper[4868]: I1201 18:10:39.946252 4868 generic.go:334] "Generic (PLEG): container finished" podID="3cc579ef-54f6-41a8-83c9-27830c9602bc" containerID="965b279256caf6c73f7fc9e2cb17f071e4fcf4fe66d7f0b24beda345c4562d91" exitCode=0 Dec 01 18:10:39 crc kubenswrapper[4868]: I1201 18:10:39.946306 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2dkl" event={"ID":"3cc579ef-54f6-41a8-83c9-27830c9602bc","Type":"ContainerDied","Data":"965b279256caf6c73f7fc9e2cb17f071e4fcf4fe66d7f0b24beda345c4562d91"} Dec 01 18:10:41 crc kubenswrapper[4868]: I1201 18:10:41.964915 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2dkl" event={"ID":"3cc579ef-54f6-41a8-83c9-27830c9602bc","Type":"ContainerStarted","Data":"2c234567f406f81b35ae2aa7bebb297ff538e3408c82618dce1608dc7f510269"} Dec 01 18:10:41 crc kubenswrapper[4868]: I1201 18:10:41.982264 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q2dkl" podStartSLOduration=3.123657631 podStartE2EDuration="5.98224244s" podCreationTimestamp="2025-12-01 18:10:36 +0000 UTC" firstStartedPulling="2025-12-01 18:10:37.925814166 +0000 UTC m=+2710.296924577" lastFinishedPulling="2025-12-01 18:10:40.784398975 +0000 UTC m=+2713.155509386" observedRunningTime="2025-12-01 18:10:41.980353488 +0000 UTC m=+2714.351463899" watchObservedRunningTime="2025-12-01 18:10:41.98224244 +0000 UTC m=+2714.353352851" Dec 01 18:10:46 crc kubenswrapper[4868]: I1201 18:10:46.723927 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:46 crc kubenswrapper[4868]: I1201 18:10:46.724525 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:46 crc kubenswrapper[4868]: I1201 18:10:46.767935 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:47 crc kubenswrapper[4868]: I1201 18:10:47.055061 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:47 crc kubenswrapper[4868]: I1201 18:10:47.101230 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q2dkl"] Dec 01 18:10:49 crc kubenswrapper[4868]: I1201 18:10:49.027079 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q2dkl" podUID="3cc579ef-54f6-41a8-83c9-27830c9602bc" containerName="registry-server" containerID="cri-o://2c234567f406f81b35ae2aa7bebb297ff538e3408c82618dce1608dc7f510269" gracePeriod=2 Dec 01 18:10:49 crc kubenswrapper[4868]: I1201 18:10:49.537369 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:49 crc kubenswrapper[4868]: I1201 18:10:49.619842 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cc579ef-54f6-41a8-83c9-27830c9602bc-catalog-content\") pod \"3cc579ef-54f6-41a8-83c9-27830c9602bc\" (UID: \"3cc579ef-54f6-41a8-83c9-27830c9602bc\") " Dec 01 18:10:49 crc kubenswrapper[4868]: I1201 18:10:49.619985 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cc579ef-54f6-41a8-83c9-27830c9602bc-utilities\") pod \"3cc579ef-54f6-41a8-83c9-27830c9602bc\" (UID: \"3cc579ef-54f6-41a8-83c9-27830c9602bc\") " Dec 01 18:10:49 crc kubenswrapper[4868]: I1201 18:10:49.620106 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjk5g\" (UniqueName: \"kubernetes.io/projected/3cc579ef-54f6-41a8-83c9-27830c9602bc-kube-api-access-fjk5g\") pod \"3cc579ef-54f6-41a8-83c9-27830c9602bc\" (UID: \"3cc579ef-54f6-41a8-83c9-27830c9602bc\") " Dec 01 18:10:49 crc kubenswrapper[4868]: I1201 18:10:49.621036 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cc579ef-54f6-41a8-83c9-27830c9602bc-utilities" (OuterVolumeSpecName: "utilities") pod "3cc579ef-54f6-41a8-83c9-27830c9602bc" (UID: "3cc579ef-54f6-41a8-83c9-27830c9602bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:10:49 crc kubenswrapper[4868]: I1201 18:10:49.628257 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cc579ef-54f6-41a8-83c9-27830c9602bc-kube-api-access-fjk5g" (OuterVolumeSpecName: "kube-api-access-fjk5g") pod "3cc579ef-54f6-41a8-83c9-27830c9602bc" (UID: "3cc579ef-54f6-41a8-83c9-27830c9602bc"). InnerVolumeSpecName "kube-api-access-fjk5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:10:49 crc kubenswrapper[4868]: I1201 18:10:49.678991 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cc579ef-54f6-41a8-83c9-27830c9602bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3cc579ef-54f6-41a8-83c9-27830c9602bc" (UID: "3cc579ef-54f6-41a8-83c9-27830c9602bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:10:49 crc kubenswrapper[4868]: I1201 18:10:49.722528 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjk5g\" (UniqueName: \"kubernetes.io/projected/3cc579ef-54f6-41a8-83c9-27830c9602bc-kube-api-access-fjk5g\") on node \"crc\" DevicePath \"\"" Dec 01 18:10:49 crc kubenswrapper[4868]: I1201 18:10:49.722562 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cc579ef-54f6-41a8-83c9-27830c9602bc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 18:10:49 crc kubenswrapper[4868]: I1201 18:10:49.722573 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cc579ef-54f6-41a8-83c9-27830c9602bc-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.041262 4868 generic.go:334] "Generic (PLEG): container finished" podID="3cc579ef-54f6-41a8-83c9-27830c9602bc" containerID="2c234567f406f81b35ae2aa7bebb297ff538e3408c82618dce1608dc7f510269" exitCode=0 Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.041307 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2dkl" event={"ID":"3cc579ef-54f6-41a8-83c9-27830c9602bc","Type":"ContainerDied","Data":"2c234567f406f81b35ae2aa7bebb297ff538e3408c82618dce1608dc7f510269"} Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.041377 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2dkl" event={"ID":"3cc579ef-54f6-41a8-83c9-27830c9602bc","Type":"ContainerDied","Data":"9029f56fbd961b5980625c9c321849e36cc4620beb23af7def1f2eab2daa952f"} Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.041399 4868 scope.go:117] "RemoveContainer" containerID="2c234567f406f81b35ae2aa7bebb297ff538e3408c82618dce1608dc7f510269" Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.042200 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2dkl" Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.062210 4868 scope.go:117] "RemoveContainer" containerID="965b279256caf6c73f7fc9e2cb17f071e4fcf4fe66d7f0b24beda345c4562d91" Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.080915 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q2dkl"] Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.089669 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q2dkl"] Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.103307 4868 scope.go:117] "RemoveContainer" containerID="e0ea49e65ca32fa4561b8c826c3ea016fcb6bb9914cd07b0cd8f0413da885d2a" Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.132601 4868 scope.go:117] "RemoveContainer" containerID="2c234567f406f81b35ae2aa7bebb297ff538e3408c82618dce1608dc7f510269" Dec 01 18:10:50 crc kubenswrapper[4868]: E1201 18:10:50.133216 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c234567f406f81b35ae2aa7bebb297ff538e3408c82618dce1608dc7f510269\": container with ID starting with 2c234567f406f81b35ae2aa7bebb297ff538e3408c82618dce1608dc7f510269 not found: ID does not exist" containerID="2c234567f406f81b35ae2aa7bebb297ff538e3408c82618dce1608dc7f510269" Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.133266 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c234567f406f81b35ae2aa7bebb297ff538e3408c82618dce1608dc7f510269"} err="failed to get container status \"2c234567f406f81b35ae2aa7bebb297ff538e3408c82618dce1608dc7f510269\": rpc error: code = NotFound desc = could not find container \"2c234567f406f81b35ae2aa7bebb297ff538e3408c82618dce1608dc7f510269\": container with ID starting with 2c234567f406f81b35ae2aa7bebb297ff538e3408c82618dce1608dc7f510269 not found: ID does not exist" Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.133293 4868 scope.go:117] "RemoveContainer" containerID="965b279256caf6c73f7fc9e2cb17f071e4fcf4fe66d7f0b24beda345c4562d91" Dec 01 18:10:50 crc kubenswrapper[4868]: E1201 18:10:50.133628 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"965b279256caf6c73f7fc9e2cb17f071e4fcf4fe66d7f0b24beda345c4562d91\": container with ID starting with 965b279256caf6c73f7fc9e2cb17f071e4fcf4fe66d7f0b24beda345c4562d91 not found: ID does not exist" containerID="965b279256caf6c73f7fc9e2cb17f071e4fcf4fe66d7f0b24beda345c4562d91" Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.133647 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"965b279256caf6c73f7fc9e2cb17f071e4fcf4fe66d7f0b24beda345c4562d91"} err="failed to get container status \"965b279256caf6c73f7fc9e2cb17f071e4fcf4fe66d7f0b24beda345c4562d91\": rpc error: code = NotFound desc = could not find container \"965b279256caf6c73f7fc9e2cb17f071e4fcf4fe66d7f0b24beda345c4562d91\": container with ID starting with 965b279256caf6c73f7fc9e2cb17f071e4fcf4fe66d7f0b24beda345c4562d91 not found: ID does not exist" Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.133664 4868 scope.go:117] "RemoveContainer" containerID="e0ea49e65ca32fa4561b8c826c3ea016fcb6bb9914cd07b0cd8f0413da885d2a" Dec 01 18:10:50 crc kubenswrapper[4868]: E1201 18:10:50.133987 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0ea49e65ca32fa4561b8c826c3ea016fcb6bb9914cd07b0cd8f0413da885d2a\": container with ID starting with e0ea49e65ca32fa4561b8c826c3ea016fcb6bb9914cd07b0cd8f0413da885d2a not found: ID does not exist" containerID="e0ea49e65ca32fa4561b8c826c3ea016fcb6bb9914cd07b0cd8f0413da885d2a" Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.134098 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0ea49e65ca32fa4561b8c826c3ea016fcb6bb9914cd07b0cd8f0413da885d2a"} err="failed to get container status \"e0ea49e65ca32fa4561b8c826c3ea016fcb6bb9914cd07b0cd8f0413da885d2a\": rpc error: code = NotFound desc = could not find container \"e0ea49e65ca32fa4561b8c826c3ea016fcb6bb9914cd07b0cd8f0413da885d2a\": container with ID starting with e0ea49e65ca32fa4561b8c826c3ea016fcb6bb9914cd07b0cd8f0413da885d2a not found: ID does not exist" Dec 01 18:10:50 crc kubenswrapper[4868]: I1201 18:10:50.185188 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cc579ef-54f6-41a8-83c9-27830c9602bc" path="/var/lib/kubelet/pods/3cc579ef-54f6-41a8-83c9-27830c9602bc/volumes" Dec 01 18:11:44 crc kubenswrapper[4868]: I1201 18:11:44.204853 4868 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-86984fb57-t5hzr" podUID="a427adaa-2060-4639-bd6d-e53b8fb00357" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 01 18:12:55 crc kubenswrapper[4868]: I1201 18:12:55.905212 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:12:55 crc kubenswrapper[4868]: I1201 18:12:55.905755 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:13:25 crc kubenswrapper[4868]: I1201 18:13:25.904162 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:13:25 crc kubenswrapper[4868]: I1201 18:13:25.904672 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:13:55 crc kubenswrapper[4868]: I1201 18:13:55.905235 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:13:55 crc kubenswrapper[4868]: I1201 18:13:55.905847 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:13:55 crc kubenswrapper[4868]: I1201 18:13:55.905913 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 18:13:55 crc kubenswrapper[4868]: I1201 18:13:55.906629 4868 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a92c978799e731b7e9c55a869165d4a8aa38fccf54030879f3328db6186e34c0"} pod="openshift-machine-config-operator/machine-config-daemon-njgpn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 18:13:55 crc kubenswrapper[4868]: I1201 18:13:55.906749 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" containerID="cri-o://a92c978799e731b7e9c55a869165d4a8aa38fccf54030879f3328db6186e34c0" gracePeriod=600 Dec 01 18:13:56 crc kubenswrapper[4868]: I1201 18:13:56.799046 4868 generic.go:334] "Generic (PLEG): container finished" podID="ec4e08bf-df56-4490-8e79-a4175c727405" containerID="a92c978799e731b7e9c55a869165d4a8aa38fccf54030879f3328db6186e34c0" exitCode=0 Dec 01 18:13:56 crc kubenswrapper[4868]: I1201 18:13:56.799123 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerDied","Data":"a92c978799e731b7e9c55a869165d4a8aa38fccf54030879f3328db6186e34c0"} Dec 01 18:13:56 crc kubenswrapper[4868]: I1201 18:13:56.800061 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6"} Dec 01 18:13:56 crc kubenswrapper[4868]: I1201 18:13:56.800091 4868 scope.go:117] "RemoveContainer" containerID="78062e97cbd766d5a4643e331fdbb29269e487b9d55f725a18c5b041495c8892" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.143676 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4"] Dec 01 18:15:00 crc kubenswrapper[4868]: E1201 18:15:00.144771 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cc579ef-54f6-41a8-83c9-27830c9602bc" containerName="registry-server" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.144791 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cc579ef-54f6-41a8-83c9-27830c9602bc" containerName="registry-server" Dec 01 18:15:00 crc kubenswrapper[4868]: E1201 18:15:00.144824 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cc579ef-54f6-41a8-83c9-27830c9602bc" containerName="extract-utilities" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.144832 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cc579ef-54f6-41a8-83c9-27830c9602bc" containerName="extract-utilities" Dec 01 18:15:00 crc kubenswrapper[4868]: E1201 18:15:00.144861 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cc579ef-54f6-41a8-83c9-27830c9602bc" containerName="extract-content" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.144869 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cc579ef-54f6-41a8-83c9-27830c9602bc" containerName="extract-content" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.145211 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cc579ef-54f6-41a8-83c9-27830c9602bc" containerName="registry-server" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.146013 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.148299 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.148340 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.155989 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4"] Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.291094 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/138fdcef-4e02-4615-9afb-a536fcb25e3d-config-volume\") pod \"collect-profiles-29410215-c46q4\" (UID: \"138fdcef-4e02-4615-9afb-a536fcb25e3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.291156 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/138fdcef-4e02-4615-9afb-a536fcb25e3d-secret-volume\") pod \"collect-profiles-29410215-c46q4\" (UID: \"138fdcef-4e02-4615-9afb-a536fcb25e3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.291605 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdw7x\" (UniqueName: \"kubernetes.io/projected/138fdcef-4e02-4615-9afb-a536fcb25e3d-kube-api-access-wdw7x\") pod \"collect-profiles-29410215-c46q4\" (UID: \"138fdcef-4e02-4615-9afb-a536fcb25e3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.393364 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdw7x\" (UniqueName: \"kubernetes.io/projected/138fdcef-4e02-4615-9afb-a536fcb25e3d-kube-api-access-wdw7x\") pod \"collect-profiles-29410215-c46q4\" (UID: \"138fdcef-4e02-4615-9afb-a536fcb25e3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.393465 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/138fdcef-4e02-4615-9afb-a536fcb25e3d-config-volume\") pod \"collect-profiles-29410215-c46q4\" (UID: \"138fdcef-4e02-4615-9afb-a536fcb25e3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.393515 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/138fdcef-4e02-4615-9afb-a536fcb25e3d-secret-volume\") pod \"collect-profiles-29410215-c46q4\" (UID: \"138fdcef-4e02-4615-9afb-a536fcb25e3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.394398 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/138fdcef-4e02-4615-9afb-a536fcb25e3d-config-volume\") pod \"collect-profiles-29410215-c46q4\" (UID: \"138fdcef-4e02-4615-9afb-a536fcb25e3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.399532 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/138fdcef-4e02-4615-9afb-a536fcb25e3d-secret-volume\") pod \"collect-profiles-29410215-c46q4\" (UID: \"138fdcef-4e02-4615-9afb-a536fcb25e3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.413643 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdw7x\" (UniqueName: \"kubernetes.io/projected/138fdcef-4e02-4615-9afb-a536fcb25e3d-kube-api-access-wdw7x\") pod \"collect-profiles-29410215-c46q4\" (UID: \"138fdcef-4e02-4615-9afb-a536fcb25e3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.474843 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" Dec 01 18:15:00 crc kubenswrapper[4868]: I1201 18:15:00.962704 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4"] Dec 01 18:15:01 crc kubenswrapper[4868]: I1201 18:15:01.872773 4868 generic.go:334] "Generic (PLEG): container finished" podID="138fdcef-4e02-4615-9afb-a536fcb25e3d" containerID="342a8d2410a3290e29044f4fda4212cf679b0cc65bdd9c6615729cb714190077" exitCode=0 Dec 01 18:15:01 crc kubenswrapper[4868]: I1201 18:15:01.872844 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" event={"ID":"138fdcef-4e02-4615-9afb-a536fcb25e3d","Type":"ContainerDied","Data":"342a8d2410a3290e29044f4fda4212cf679b0cc65bdd9c6615729cb714190077"} Dec 01 18:15:01 crc kubenswrapper[4868]: I1201 18:15:01.873078 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" event={"ID":"138fdcef-4e02-4615-9afb-a536fcb25e3d","Type":"ContainerStarted","Data":"7707d1430cbb00f881692c044a6b6088143a4d734ff6b27fa686b8e7a55ee999"} Dec 01 18:15:03 crc kubenswrapper[4868]: I1201 18:15:03.241046 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" Dec 01 18:15:03 crc kubenswrapper[4868]: I1201 18:15:03.343796 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdw7x\" (UniqueName: \"kubernetes.io/projected/138fdcef-4e02-4615-9afb-a536fcb25e3d-kube-api-access-wdw7x\") pod \"138fdcef-4e02-4615-9afb-a536fcb25e3d\" (UID: \"138fdcef-4e02-4615-9afb-a536fcb25e3d\") " Dec 01 18:15:03 crc kubenswrapper[4868]: I1201 18:15:03.343932 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/138fdcef-4e02-4615-9afb-a536fcb25e3d-config-volume\") pod \"138fdcef-4e02-4615-9afb-a536fcb25e3d\" (UID: \"138fdcef-4e02-4615-9afb-a536fcb25e3d\") " Dec 01 18:15:03 crc kubenswrapper[4868]: I1201 18:15:03.344000 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/138fdcef-4e02-4615-9afb-a536fcb25e3d-secret-volume\") pod \"138fdcef-4e02-4615-9afb-a536fcb25e3d\" (UID: \"138fdcef-4e02-4615-9afb-a536fcb25e3d\") " Dec 01 18:15:03 crc kubenswrapper[4868]: I1201 18:15:03.344783 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/138fdcef-4e02-4615-9afb-a536fcb25e3d-config-volume" (OuterVolumeSpecName: "config-volume") pod "138fdcef-4e02-4615-9afb-a536fcb25e3d" (UID: "138fdcef-4e02-4615-9afb-a536fcb25e3d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 18:15:03 crc kubenswrapper[4868]: I1201 18:15:03.350964 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/138fdcef-4e02-4615-9afb-a536fcb25e3d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "138fdcef-4e02-4615-9afb-a536fcb25e3d" (UID: "138fdcef-4e02-4615-9afb-a536fcb25e3d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:15:03 crc kubenswrapper[4868]: I1201 18:15:03.354024 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/138fdcef-4e02-4615-9afb-a536fcb25e3d-kube-api-access-wdw7x" (OuterVolumeSpecName: "kube-api-access-wdw7x") pod "138fdcef-4e02-4615-9afb-a536fcb25e3d" (UID: "138fdcef-4e02-4615-9afb-a536fcb25e3d"). InnerVolumeSpecName "kube-api-access-wdw7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:15:03 crc kubenswrapper[4868]: I1201 18:15:03.446801 4868 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/138fdcef-4e02-4615-9afb-a536fcb25e3d-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 18:15:03 crc kubenswrapper[4868]: I1201 18:15:03.446858 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdw7x\" (UniqueName: \"kubernetes.io/projected/138fdcef-4e02-4615-9afb-a536fcb25e3d-kube-api-access-wdw7x\") on node \"crc\" DevicePath \"\"" Dec 01 18:15:03 crc kubenswrapper[4868]: I1201 18:15:03.446868 4868 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/138fdcef-4e02-4615-9afb-a536fcb25e3d-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 18:15:03 crc kubenswrapper[4868]: I1201 18:15:03.889473 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" event={"ID":"138fdcef-4e02-4615-9afb-a536fcb25e3d","Type":"ContainerDied","Data":"7707d1430cbb00f881692c044a6b6088143a4d734ff6b27fa686b8e7a55ee999"} Dec 01 18:15:03 crc kubenswrapper[4868]: I1201 18:15:03.889517 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7707d1430cbb00f881692c044a6b6088143a4d734ff6b27fa686b8e7a55ee999" Dec 01 18:15:03 crc kubenswrapper[4868]: I1201 18:15:03.889606 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410215-c46q4" Dec 01 18:15:04 crc kubenswrapper[4868]: I1201 18:15:04.321298 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s"] Dec 01 18:15:04 crc kubenswrapper[4868]: I1201 18:15:04.333047 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410170-4w22s"] Dec 01 18:15:06 crc kubenswrapper[4868]: I1201 18:15:06.186187 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee6b0713-849b-4108-8fa0-53739ec95458" path="/var/lib/kubelet/pods/ee6b0713-849b-4108-8fa0-53739ec95458/volumes" Dec 01 18:15:41 crc kubenswrapper[4868]: I1201 18:15:41.684473 4868 scope.go:117] "RemoveContainer" containerID="8f5fa2feac10e3963105ad4104d3c6ac4c849489a67233fd901c985ec658978c" Dec 01 18:16:25 crc kubenswrapper[4868]: I1201 18:16:25.905381 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:16:25 crc kubenswrapper[4868]: I1201 18:16:25.906192 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:16:55 crc kubenswrapper[4868]: I1201 18:16:55.904933 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:16:55 crc kubenswrapper[4868]: I1201 18:16:55.905470 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:17:09 crc kubenswrapper[4868]: I1201 18:17:09.409992 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wwcp6"] Dec 01 18:17:09 crc kubenswrapper[4868]: E1201 18:17:09.410928 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="138fdcef-4e02-4615-9afb-a536fcb25e3d" containerName="collect-profiles" Dec 01 18:17:09 crc kubenswrapper[4868]: I1201 18:17:09.410962 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="138fdcef-4e02-4615-9afb-a536fcb25e3d" containerName="collect-profiles" Dec 01 18:17:09 crc kubenswrapper[4868]: I1201 18:17:09.411131 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="138fdcef-4e02-4615-9afb-a536fcb25e3d" containerName="collect-profiles" Dec 01 18:17:09 crc kubenswrapper[4868]: I1201 18:17:09.412393 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:09 crc kubenswrapper[4868]: I1201 18:17:09.433151 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wwcp6"] Dec 01 18:17:09 crc kubenswrapper[4868]: I1201 18:17:09.515751 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdr5m\" (UniqueName: \"kubernetes.io/projected/9af1bcb6-23cd-41a4-a98d-ce5360928b43-kube-api-access-bdr5m\") pod \"community-operators-wwcp6\" (UID: \"9af1bcb6-23cd-41a4-a98d-ce5360928b43\") " pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:09 crc kubenswrapper[4868]: I1201 18:17:09.515803 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9af1bcb6-23cd-41a4-a98d-ce5360928b43-utilities\") pod \"community-operators-wwcp6\" (UID: \"9af1bcb6-23cd-41a4-a98d-ce5360928b43\") " pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:09 crc kubenswrapper[4868]: I1201 18:17:09.515933 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9af1bcb6-23cd-41a4-a98d-ce5360928b43-catalog-content\") pod \"community-operators-wwcp6\" (UID: \"9af1bcb6-23cd-41a4-a98d-ce5360928b43\") " pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:09 crc kubenswrapper[4868]: I1201 18:17:09.617779 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9af1bcb6-23cd-41a4-a98d-ce5360928b43-catalog-content\") pod \"community-operators-wwcp6\" (UID: \"9af1bcb6-23cd-41a4-a98d-ce5360928b43\") " pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:09 crc kubenswrapper[4868]: I1201 18:17:09.618310 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdr5m\" (UniqueName: \"kubernetes.io/projected/9af1bcb6-23cd-41a4-a98d-ce5360928b43-kube-api-access-bdr5m\") pod \"community-operators-wwcp6\" (UID: \"9af1bcb6-23cd-41a4-a98d-ce5360928b43\") " pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:09 crc kubenswrapper[4868]: I1201 18:17:09.618335 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9af1bcb6-23cd-41a4-a98d-ce5360928b43-utilities\") pod \"community-operators-wwcp6\" (UID: \"9af1bcb6-23cd-41a4-a98d-ce5360928b43\") " pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:09 crc kubenswrapper[4868]: I1201 18:17:09.618987 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9af1bcb6-23cd-41a4-a98d-ce5360928b43-catalog-content\") pod \"community-operators-wwcp6\" (UID: \"9af1bcb6-23cd-41a4-a98d-ce5360928b43\") " pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:09 crc kubenswrapper[4868]: I1201 18:17:09.619083 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9af1bcb6-23cd-41a4-a98d-ce5360928b43-utilities\") pod \"community-operators-wwcp6\" (UID: \"9af1bcb6-23cd-41a4-a98d-ce5360928b43\") " pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:09 crc kubenswrapper[4868]: I1201 18:17:09.642727 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdr5m\" (UniqueName: \"kubernetes.io/projected/9af1bcb6-23cd-41a4-a98d-ce5360928b43-kube-api-access-bdr5m\") pod \"community-operators-wwcp6\" (UID: \"9af1bcb6-23cd-41a4-a98d-ce5360928b43\") " pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:09 crc kubenswrapper[4868]: I1201 18:17:09.734595 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:10 crc kubenswrapper[4868]: I1201 18:17:10.301132 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wwcp6"] Dec 01 18:17:10 crc kubenswrapper[4868]: I1201 18:17:10.332610 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wwcp6" event={"ID":"9af1bcb6-23cd-41a4-a98d-ce5360928b43","Type":"ContainerStarted","Data":"fe23e9c4350dddf780ac2ac9cead8e06f1565029f24bcb1298bcc72c387a9ce7"} Dec 01 18:17:11 crc kubenswrapper[4868]: I1201 18:17:11.344237 4868 generic.go:334] "Generic (PLEG): container finished" podID="9af1bcb6-23cd-41a4-a98d-ce5360928b43" containerID="e3c75bd92e00573f144fa2c1ba053a044d5d9cba77a81b3dc14b138cc96093ba" exitCode=0 Dec 01 18:17:11 crc kubenswrapper[4868]: I1201 18:17:11.344310 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wwcp6" event={"ID":"9af1bcb6-23cd-41a4-a98d-ce5360928b43","Type":"ContainerDied","Data":"e3c75bd92e00573f144fa2c1ba053a044d5d9cba77a81b3dc14b138cc96093ba"} Dec 01 18:17:11 crc kubenswrapper[4868]: I1201 18:17:11.348551 4868 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.012706 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-njnrv"] Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.015503 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.050470 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-njnrv"] Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.093545 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60f2cfb2-9ecf-4666-b53b-681a6463e45b-catalog-content\") pod \"redhat-marketplace-njnrv\" (UID: \"60f2cfb2-9ecf-4666-b53b-681a6463e45b\") " pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.093693 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xgvv\" (UniqueName: \"kubernetes.io/projected/60f2cfb2-9ecf-4666-b53b-681a6463e45b-kube-api-access-5xgvv\") pod \"redhat-marketplace-njnrv\" (UID: \"60f2cfb2-9ecf-4666-b53b-681a6463e45b\") " pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.093858 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60f2cfb2-9ecf-4666-b53b-681a6463e45b-utilities\") pod \"redhat-marketplace-njnrv\" (UID: \"60f2cfb2-9ecf-4666-b53b-681a6463e45b\") " pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.196022 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60f2cfb2-9ecf-4666-b53b-681a6463e45b-utilities\") pod \"redhat-marketplace-njnrv\" (UID: \"60f2cfb2-9ecf-4666-b53b-681a6463e45b\") " pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.196344 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60f2cfb2-9ecf-4666-b53b-681a6463e45b-catalog-content\") pod \"redhat-marketplace-njnrv\" (UID: \"60f2cfb2-9ecf-4666-b53b-681a6463e45b\") " pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.196504 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xgvv\" (UniqueName: \"kubernetes.io/projected/60f2cfb2-9ecf-4666-b53b-681a6463e45b-kube-api-access-5xgvv\") pod \"redhat-marketplace-njnrv\" (UID: \"60f2cfb2-9ecf-4666-b53b-681a6463e45b\") " pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.196691 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60f2cfb2-9ecf-4666-b53b-681a6463e45b-utilities\") pod \"redhat-marketplace-njnrv\" (UID: \"60f2cfb2-9ecf-4666-b53b-681a6463e45b\") " pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.196803 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60f2cfb2-9ecf-4666-b53b-681a6463e45b-catalog-content\") pod \"redhat-marketplace-njnrv\" (UID: \"60f2cfb2-9ecf-4666-b53b-681a6463e45b\") " pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.217426 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xgvv\" (UniqueName: \"kubernetes.io/projected/60f2cfb2-9ecf-4666-b53b-681a6463e45b-kube-api-access-5xgvv\") pod \"redhat-marketplace-njnrv\" (UID: \"60f2cfb2-9ecf-4666-b53b-681a6463e45b\") " pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.356844 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.614001 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8trm7"] Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.616668 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.639019 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8trm7"] Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.708279 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c1bdc3-084e-4f96-874b-95c1b143b485-utilities\") pod \"redhat-operators-8trm7\" (UID: \"30c1bdc3-084e-4f96-874b-95c1b143b485\") " pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.708351 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xrh4\" (UniqueName: \"kubernetes.io/projected/30c1bdc3-084e-4f96-874b-95c1b143b485-kube-api-access-5xrh4\") pod \"redhat-operators-8trm7\" (UID: \"30c1bdc3-084e-4f96-874b-95c1b143b485\") " pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.708390 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c1bdc3-084e-4f96-874b-95c1b143b485-catalog-content\") pod \"redhat-operators-8trm7\" (UID: \"30c1bdc3-084e-4f96-874b-95c1b143b485\") " pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.810343 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c1bdc3-084e-4f96-874b-95c1b143b485-utilities\") pod \"redhat-operators-8trm7\" (UID: \"30c1bdc3-084e-4f96-874b-95c1b143b485\") " pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.810430 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xrh4\" (UniqueName: \"kubernetes.io/projected/30c1bdc3-084e-4f96-874b-95c1b143b485-kube-api-access-5xrh4\") pod \"redhat-operators-8trm7\" (UID: \"30c1bdc3-084e-4f96-874b-95c1b143b485\") " pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.810471 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c1bdc3-084e-4f96-874b-95c1b143b485-catalog-content\") pod \"redhat-operators-8trm7\" (UID: \"30c1bdc3-084e-4f96-874b-95c1b143b485\") " pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.810826 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c1bdc3-084e-4f96-874b-95c1b143b485-utilities\") pod \"redhat-operators-8trm7\" (UID: \"30c1bdc3-084e-4f96-874b-95c1b143b485\") " pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.810862 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c1bdc3-084e-4f96-874b-95c1b143b485-catalog-content\") pod \"redhat-operators-8trm7\" (UID: \"30c1bdc3-084e-4f96-874b-95c1b143b485\") " pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.833121 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xrh4\" (UniqueName: \"kubernetes.io/projected/30c1bdc3-084e-4f96-874b-95c1b143b485-kube-api-access-5xrh4\") pod \"redhat-operators-8trm7\" (UID: \"30c1bdc3-084e-4f96-874b-95c1b143b485\") " pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.903134 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-njnrv"] Dec 01 18:17:12 crc kubenswrapper[4868]: W1201 18:17:12.917686 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60f2cfb2_9ecf_4666_b53b_681a6463e45b.slice/crio-ed8bfdc8ce9df33df54848847198ec920857d729eacb5732e0cc9b5f3cccecea WatchSource:0}: Error finding container ed8bfdc8ce9df33df54848847198ec920857d729eacb5732e0cc9b5f3cccecea: Status 404 returned error can't find the container with id ed8bfdc8ce9df33df54848847198ec920857d729eacb5732e0cc9b5f3cccecea Dec 01 18:17:12 crc kubenswrapper[4868]: I1201 18:17:12.945779 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:13 crc kubenswrapper[4868]: I1201 18:17:13.367751 4868 generic.go:334] "Generic (PLEG): container finished" podID="9af1bcb6-23cd-41a4-a98d-ce5360928b43" containerID="5f0d6e123c3344427c4eb5c194d73105d629469ad9427f15e116c0eb0452fec1" exitCode=0 Dec 01 18:17:13 crc kubenswrapper[4868]: I1201 18:17:13.367887 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wwcp6" event={"ID":"9af1bcb6-23cd-41a4-a98d-ce5360928b43","Type":"ContainerDied","Data":"5f0d6e123c3344427c4eb5c194d73105d629469ad9427f15e116c0eb0452fec1"} Dec 01 18:17:13 crc kubenswrapper[4868]: I1201 18:17:13.373614 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njnrv" event={"ID":"60f2cfb2-9ecf-4666-b53b-681a6463e45b","Type":"ContainerStarted","Data":"c8d5f87a8c0d7c53ed37cde41c00cff6e2e2e32e559d060b519bd1f0a50a2630"} Dec 01 18:17:13 crc kubenswrapper[4868]: I1201 18:17:13.373726 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njnrv" event={"ID":"60f2cfb2-9ecf-4666-b53b-681a6463e45b","Type":"ContainerStarted","Data":"ed8bfdc8ce9df33df54848847198ec920857d729eacb5732e0cc9b5f3cccecea"} Dec 01 18:17:13 crc kubenswrapper[4868]: I1201 18:17:13.456435 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8trm7"] Dec 01 18:17:13 crc kubenswrapper[4868]: W1201 18:17:13.470442 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30c1bdc3_084e_4f96_874b_95c1b143b485.slice/crio-27a59d8ee12e750b1edb0d6ea3c52a12fcbedc3cc18cd9fb76bdff197da3b4df WatchSource:0}: Error finding container 27a59d8ee12e750b1edb0d6ea3c52a12fcbedc3cc18cd9fb76bdff197da3b4df: Status 404 returned error can't find the container with id 27a59d8ee12e750b1edb0d6ea3c52a12fcbedc3cc18cd9fb76bdff197da3b4df Dec 01 18:17:14 crc kubenswrapper[4868]: I1201 18:17:14.390900 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wwcp6" event={"ID":"9af1bcb6-23cd-41a4-a98d-ce5360928b43","Type":"ContainerStarted","Data":"291503256a24c9361b94a48bb468f7a22eb93773e030878cdf3221d52e7e7adc"} Dec 01 18:17:14 crc kubenswrapper[4868]: I1201 18:17:14.394415 4868 generic.go:334] "Generic (PLEG): container finished" podID="30c1bdc3-084e-4f96-874b-95c1b143b485" containerID="f105bdabee44e7954ff513192f754af192ef7a2fe810236288ea54a66f2ded4d" exitCode=0 Dec 01 18:17:14 crc kubenswrapper[4868]: I1201 18:17:14.394491 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8trm7" event={"ID":"30c1bdc3-084e-4f96-874b-95c1b143b485","Type":"ContainerDied","Data":"f105bdabee44e7954ff513192f754af192ef7a2fe810236288ea54a66f2ded4d"} Dec 01 18:17:14 crc kubenswrapper[4868]: I1201 18:17:14.394516 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8trm7" event={"ID":"30c1bdc3-084e-4f96-874b-95c1b143b485","Type":"ContainerStarted","Data":"27a59d8ee12e750b1edb0d6ea3c52a12fcbedc3cc18cd9fb76bdff197da3b4df"} Dec 01 18:17:14 crc kubenswrapper[4868]: I1201 18:17:14.396963 4868 generic.go:334] "Generic (PLEG): container finished" podID="60f2cfb2-9ecf-4666-b53b-681a6463e45b" containerID="c8d5f87a8c0d7c53ed37cde41c00cff6e2e2e32e559d060b519bd1f0a50a2630" exitCode=0 Dec 01 18:17:14 crc kubenswrapper[4868]: I1201 18:17:14.397011 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njnrv" event={"ID":"60f2cfb2-9ecf-4666-b53b-681a6463e45b","Type":"ContainerDied","Data":"c8d5f87a8c0d7c53ed37cde41c00cff6e2e2e32e559d060b519bd1f0a50a2630"} Dec 01 18:17:14 crc kubenswrapper[4868]: I1201 18:17:14.417854 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wwcp6" podStartSLOduration=2.850840688 podStartE2EDuration="5.417833153s" podCreationTimestamp="2025-12-01 18:17:09 +0000 UTC" firstStartedPulling="2025-12-01 18:17:11.347831163 +0000 UTC m=+3103.718941624" lastFinishedPulling="2025-12-01 18:17:13.914823678 +0000 UTC m=+3106.285934089" observedRunningTime="2025-12-01 18:17:14.414704449 +0000 UTC m=+3106.785814860" watchObservedRunningTime="2025-12-01 18:17:14.417833153 +0000 UTC m=+3106.788943564" Dec 01 18:17:15 crc kubenswrapper[4868]: I1201 18:17:15.408897 4868 generic.go:334] "Generic (PLEG): container finished" podID="60f2cfb2-9ecf-4666-b53b-681a6463e45b" containerID="f01df34e17a0124c8660d63677a411b0beb4f50530b4e1a0904ececde3fa46e4" exitCode=0 Dec 01 18:17:15 crc kubenswrapper[4868]: I1201 18:17:15.408982 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njnrv" event={"ID":"60f2cfb2-9ecf-4666-b53b-681a6463e45b","Type":"ContainerDied","Data":"f01df34e17a0124c8660d63677a411b0beb4f50530b4e1a0904ececde3fa46e4"} Dec 01 18:17:16 crc kubenswrapper[4868]: I1201 18:17:16.421104 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8trm7" event={"ID":"30c1bdc3-084e-4f96-874b-95c1b143b485","Type":"ContainerStarted","Data":"cb7a9c00e4697f9c8092a32bcc7446d43e1ffc498b8d9170182407934dd32723"} Dec 01 18:17:16 crc kubenswrapper[4868]: I1201 18:17:16.423313 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njnrv" event={"ID":"60f2cfb2-9ecf-4666-b53b-681a6463e45b","Type":"ContainerStarted","Data":"3c98af99303d7f70ce714780f2b5a6d0e93c3a4ed4f61e381671b9a223cec901"} Dec 01 18:17:16 crc kubenswrapper[4868]: I1201 18:17:16.463189 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-njnrv" podStartSLOduration=2.7236717390000003 podStartE2EDuration="5.463164678s" podCreationTimestamp="2025-12-01 18:17:11 +0000 UTC" firstStartedPulling="2025-12-01 18:17:13.376091723 +0000 UTC m=+3105.747202134" lastFinishedPulling="2025-12-01 18:17:16.115584662 +0000 UTC m=+3108.486695073" observedRunningTime="2025-12-01 18:17:16.45884358 +0000 UTC m=+3108.829953991" watchObservedRunningTime="2025-12-01 18:17:16.463164678 +0000 UTC m=+3108.834275089" Dec 01 18:17:18 crc kubenswrapper[4868]: I1201 18:17:18.451780 4868 generic.go:334] "Generic (PLEG): container finished" podID="30c1bdc3-084e-4f96-874b-95c1b143b485" containerID="cb7a9c00e4697f9c8092a32bcc7446d43e1ffc498b8d9170182407934dd32723" exitCode=0 Dec 01 18:17:18 crc kubenswrapper[4868]: I1201 18:17:18.451873 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8trm7" event={"ID":"30c1bdc3-084e-4f96-874b-95c1b143b485","Type":"ContainerDied","Data":"cb7a9c00e4697f9c8092a32bcc7446d43e1ffc498b8d9170182407934dd32723"} Dec 01 18:17:19 crc kubenswrapper[4868]: I1201 18:17:19.735297 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:19 crc kubenswrapper[4868]: I1201 18:17:19.735704 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:20 crc kubenswrapper[4868]: I1201 18:17:20.471853 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8trm7" event={"ID":"30c1bdc3-084e-4f96-874b-95c1b143b485","Type":"ContainerStarted","Data":"b9b999de370b6d397a37f23c6a0234a545fc002141e10c5ec8f28e7f614a234a"} Dec 01 18:17:20 crc kubenswrapper[4868]: I1201 18:17:20.501765 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8trm7" podStartSLOduration=3.190479739 podStartE2EDuration="8.501743883s" podCreationTimestamp="2025-12-01 18:17:12 +0000 UTC" firstStartedPulling="2025-12-01 18:17:14.396804392 +0000 UTC m=+3106.767914803" lastFinishedPulling="2025-12-01 18:17:19.708068536 +0000 UTC m=+3112.079178947" observedRunningTime="2025-12-01 18:17:20.494587988 +0000 UTC m=+3112.865698399" watchObservedRunningTime="2025-12-01 18:17:20.501743883 +0000 UTC m=+3112.872854294" Dec 01 18:17:20 crc kubenswrapper[4868]: I1201 18:17:20.793998 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-wwcp6" podUID="9af1bcb6-23cd-41a4-a98d-ce5360928b43" containerName="registry-server" probeResult="failure" output=< Dec 01 18:17:20 crc kubenswrapper[4868]: timeout: failed to connect service ":50051" within 1s Dec 01 18:17:20 crc kubenswrapper[4868]: > Dec 01 18:17:22 crc kubenswrapper[4868]: I1201 18:17:22.358078 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:22 crc kubenswrapper[4868]: I1201 18:17:22.361099 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:22 crc kubenswrapper[4868]: I1201 18:17:22.451393 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:22 crc kubenswrapper[4868]: I1201 18:17:22.554201 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:22 crc kubenswrapper[4868]: I1201 18:17:22.946786 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:22 crc kubenswrapper[4868]: I1201 18:17:22.946850 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:23 crc kubenswrapper[4868]: I1201 18:17:23.601707 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-njnrv"] Dec 01 18:17:23 crc kubenswrapper[4868]: I1201 18:17:23.995539 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8trm7" podUID="30c1bdc3-084e-4f96-874b-95c1b143b485" containerName="registry-server" probeResult="failure" output=< Dec 01 18:17:23 crc kubenswrapper[4868]: timeout: failed to connect service ":50051" within 1s Dec 01 18:17:23 crc kubenswrapper[4868]: > Dec 01 18:17:25 crc kubenswrapper[4868]: I1201 18:17:25.516836 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-njnrv" podUID="60f2cfb2-9ecf-4666-b53b-681a6463e45b" containerName="registry-server" containerID="cri-o://3c98af99303d7f70ce714780f2b5a6d0e93c3a4ed4f61e381671b9a223cec901" gracePeriod=2 Dec 01 18:17:25 crc kubenswrapper[4868]: I1201 18:17:25.904915 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:17:25 crc kubenswrapper[4868]: I1201 18:17:25.904985 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:17:25 crc kubenswrapper[4868]: I1201 18:17:25.905028 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 18:17:25 crc kubenswrapper[4868]: I1201 18:17:25.909738 4868 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6"} pod="openshift-machine-config-operator/machine-config-daemon-njgpn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 18:17:25 crc kubenswrapper[4868]: I1201 18:17:25.909818 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" containerID="cri-o://9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" gracePeriod=600 Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.014682 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:26 crc kubenswrapper[4868]: E1201 18:17:26.043556 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.104047 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xgvv\" (UniqueName: \"kubernetes.io/projected/60f2cfb2-9ecf-4666-b53b-681a6463e45b-kube-api-access-5xgvv\") pod \"60f2cfb2-9ecf-4666-b53b-681a6463e45b\" (UID: \"60f2cfb2-9ecf-4666-b53b-681a6463e45b\") " Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.104113 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60f2cfb2-9ecf-4666-b53b-681a6463e45b-utilities\") pod \"60f2cfb2-9ecf-4666-b53b-681a6463e45b\" (UID: \"60f2cfb2-9ecf-4666-b53b-681a6463e45b\") " Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.104208 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60f2cfb2-9ecf-4666-b53b-681a6463e45b-catalog-content\") pod \"60f2cfb2-9ecf-4666-b53b-681a6463e45b\" (UID: \"60f2cfb2-9ecf-4666-b53b-681a6463e45b\") " Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.104800 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60f2cfb2-9ecf-4666-b53b-681a6463e45b-utilities" (OuterVolumeSpecName: "utilities") pod "60f2cfb2-9ecf-4666-b53b-681a6463e45b" (UID: "60f2cfb2-9ecf-4666-b53b-681a6463e45b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.105116 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60f2cfb2-9ecf-4666-b53b-681a6463e45b-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.110029 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60f2cfb2-9ecf-4666-b53b-681a6463e45b-kube-api-access-5xgvv" (OuterVolumeSpecName: "kube-api-access-5xgvv") pod "60f2cfb2-9ecf-4666-b53b-681a6463e45b" (UID: "60f2cfb2-9ecf-4666-b53b-681a6463e45b"). InnerVolumeSpecName "kube-api-access-5xgvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.120822 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60f2cfb2-9ecf-4666-b53b-681a6463e45b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "60f2cfb2-9ecf-4666-b53b-681a6463e45b" (UID: "60f2cfb2-9ecf-4666-b53b-681a6463e45b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.206450 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60f2cfb2-9ecf-4666-b53b-681a6463e45b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.206483 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xgvv\" (UniqueName: \"kubernetes.io/projected/60f2cfb2-9ecf-4666-b53b-681a6463e45b-kube-api-access-5xgvv\") on node \"crc\" DevicePath \"\"" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.527707 4868 generic.go:334] "Generic (PLEG): container finished" podID="ec4e08bf-df56-4490-8e79-a4175c727405" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" exitCode=0 Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.527778 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerDied","Data":"9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6"} Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.527810 4868 scope.go:117] "RemoveContainer" containerID="a92c978799e731b7e9c55a869165d4a8aa38fccf54030879f3328db6186e34c0" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.528606 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:17:26 crc kubenswrapper[4868]: E1201 18:17:26.528997 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.530815 4868 generic.go:334] "Generic (PLEG): container finished" podID="60f2cfb2-9ecf-4666-b53b-681a6463e45b" containerID="3c98af99303d7f70ce714780f2b5a6d0e93c3a4ed4f61e381671b9a223cec901" exitCode=0 Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.530839 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njnrv" event={"ID":"60f2cfb2-9ecf-4666-b53b-681a6463e45b","Type":"ContainerDied","Data":"3c98af99303d7f70ce714780f2b5a6d0e93c3a4ed4f61e381671b9a223cec901"} Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.530855 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-njnrv" event={"ID":"60f2cfb2-9ecf-4666-b53b-681a6463e45b","Type":"ContainerDied","Data":"ed8bfdc8ce9df33df54848847198ec920857d729eacb5732e0cc9b5f3cccecea"} Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.530911 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-njnrv" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.562702 4868 scope.go:117] "RemoveContainer" containerID="3c98af99303d7f70ce714780f2b5a6d0e93c3a4ed4f61e381671b9a223cec901" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.572256 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-njnrv"] Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.582882 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-njnrv"] Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.586090 4868 scope.go:117] "RemoveContainer" containerID="f01df34e17a0124c8660d63677a411b0beb4f50530b4e1a0904ececde3fa46e4" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.616032 4868 scope.go:117] "RemoveContainer" containerID="c8d5f87a8c0d7c53ed37cde41c00cff6e2e2e32e559d060b519bd1f0a50a2630" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.644141 4868 scope.go:117] "RemoveContainer" containerID="3c98af99303d7f70ce714780f2b5a6d0e93c3a4ed4f61e381671b9a223cec901" Dec 01 18:17:26 crc kubenswrapper[4868]: E1201 18:17:26.644545 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c98af99303d7f70ce714780f2b5a6d0e93c3a4ed4f61e381671b9a223cec901\": container with ID starting with 3c98af99303d7f70ce714780f2b5a6d0e93c3a4ed4f61e381671b9a223cec901 not found: ID does not exist" containerID="3c98af99303d7f70ce714780f2b5a6d0e93c3a4ed4f61e381671b9a223cec901" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.644583 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c98af99303d7f70ce714780f2b5a6d0e93c3a4ed4f61e381671b9a223cec901"} err="failed to get container status \"3c98af99303d7f70ce714780f2b5a6d0e93c3a4ed4f61e381671b9a223cec901\": rpc error: code = NotFound desc = could not find container \"3c98af99303d7f70ce714780f2b5a6d0e93c3a4ed4f61e381671b9a223cec901\": container with ID starting with 3c98af99303d7f70ce714780f2b5a6d0e93c3a4ed4f61e381671b9a223cec901 not found: ID does not exist" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.644610 4868 scope.go:117] "RemoveContainer" containerID="f01df34e17a0124c8660d63677a411b0beb4f50530b4e1a0904ececde3fa46e4" Dec 01 18:17:26 crc kubenswrapper[4868]: E1201 18:17:26.644970 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f01df34e17a0124c8660d63677a411b0beb4f50530b4e1a0904ececde3fa46e4\": container with ID starting with f01df34e17a0124c8660d63677a411b0beb4f50530b4e1a0904ececde3fa46e4 not found: ID does not exist" containerID="f01df34e17a0124c8660d63677a411b0beb4f50530b4e1a0904ececde3fa46e4" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.645000 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f01df34e17a0124c8660d63677a411b0beb4f50530b4e1a0904ececde3fa46e4"} err="failed to get container status \"f01df34e17a0124c8660d63677a411b0beb4f50530b4e1a0904ececde3fa46e4\": rpc error: code = NotFound desc = could not find container \"f01df34e17a0124c8660d63677a411b0beb4f50530b4e1a0904ececde3fa46e4\": container with ID starting with f01df34e17a0124c8660d63677a411b0beb4f50530b4e1a0904ececde3fa46e4 not found: ID does not exist" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.645025 4868 scope.go:117] "RemoveContainer" containerID="c8d5f87a8c0d7c53ed37cde41c00cff6e2e2e32e559d060b519bd1f0a50a2630" Dec 01 18:17:26 crc kubenswrapper[4868]: E1201 18:17:26.645360 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8d5f87a8c0d7c53ed37cde41c00cff6e2e2e32e559d060b519bd1f0a50a2630\": container with ID starting with c8d5f87a8c0d7c53ed37cde41c00cff6e2e2e32e559d060b519bd1f0a50a2630 not found: ID does not exist" containerID="c8d5f87a8c0d7c53ed37cde41c00cff6e2e2e32e559d060b519bd1f0a50a2630" Dec 01 18:17:26 crc kubenswrapper[4868]: I1201 18:17:26.645389 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8d5f87a8c0d7c53ed37cde41c00cff6e2e2e32e559d060b519bd1f0a50a2630"} err="failed to get container status \"c8d5f87a8c0d7c53ed37cde41c00cff6e2e2e32e559d060b519bd1f0a50a2630\": rpc error: code = NotFound desc = could not find container \"c8d5f87a8c0d7c53ed37cde41c00cff6e2e2e32e559d060b519bd1f0a50a2630\": container with ID starting with c8d5f87a8c0d7c53ed37cde41c00cff6e2e2e32e559d060b519bd1f0a50a2630 not found: ID does not exist" Dec 01 18:17:28 crc kubenswrapper[4868]: I1201 18:17:28.186038 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60f2cfb2-9ecf-4666-b53b-681a6463e45b" path="/var/lib/kubelet/pods/60f2cfb2-9ecf-4666-b53b-681a6463e45b/volumes" Dec 01 18:17:29 crc kubenswrapper[4868]: I1201 18:17:29.822162 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:29 crc kubenswrapper[4868]: I1201 18:17:29.911264 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:30 crc kubenswrapper[4868]: I1201 18:17:30.070645 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wwcp6"] Dec 01 18:17:31 crc kubenswrapper[4868]: I1201 18:17:31.582389 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wwcp6" podUID="9af1bcb6-23cd-41a4-a98d-ce5360928b43" containerName="registry-server" containerID="cri-o://291503256a24c9361b94a48bb468f7a22eb93773e030878cdf3221d52e7e7adc" gracePeriod=2 Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.081273 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.221256 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdr5m\" (UniqueName: \"kubernetes.io/projected/9af1bcb6-23cd-41a4-a98d-ce5360928b43-kube-api-access-bdr5m\") pod \"9af1bcb6-23cd-41a4-a98d-ce5360928b43\" (UID: \"9af1bcb6-23cd-41a4-a98d-ce5360928b43\") " Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.221350 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9af1bcb6-23cd-41a4-a98d-ce5360928b43-catalog-content\") pod \"9af1bcb6-23cd-41a4-a98d-ce5360928b43\" (UID: \"9af1bcb6-23cd-41a4-a98d-ce5360928b43\") " Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.221384 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9af1bcb6-23cd-41a4-a98d-ce5360928b43-utilities\") pod \"9af1bcb6-23cd-41a4-a98d-ce5360928b43\" (UID: \"9af1bcb6-23cd-41a4-a98d-ce5360928b43\") " Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.222215 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9af1bcb6-23cd-41a4-a98d-ce5360928b43-utilities" (OuterVolumeSpecName: "utilities") pod "9af1bcb6-23cd-41a4-a98d-ce5360928b43" (UID: "9af1bcb6-23cd-41a4-a98d-ce5360928b43"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.242122 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9af1bcb6-23cd-41a4-a98d-ce5360928b43-kube-api-access-bdr5m" (OuterVolumeSpecName: "kube-api-access-bdr5m") pod "9af1bcb6-23cd-41a4-a98d-ce5360928b43" (UID: "9af1bcb6-23cd-41a4-a98d-ce5360928b43"). InnerVolumeSpecName "kube-api-access-bdr5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.268386 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9af1bcb6-23cd-41a4-a98d-ce5360928b43-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9af1bcb6-23cd-41a4-a98d-ce5360928b43" (UID: "9af1bcb6-23cd-41a4-a98d-ce5360928b43"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.323425 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdr5m\" (UniqueName: \"kubernetes.io/projected/9af1bcb6-23cd-41a4-a98d-ce5360928b43-kube-api-access-bdr5m\") on node \"crc\" DevicePath \"\"" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.323463 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9af1bcb6-23cd-41a4-a98d-ce5360928b43-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.323473 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9af1bcb6-23cd-41a4-a98d-ce5360928b43-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.595683 4868 generic.go:334] "Generic (PLEG): container finished" podID="9af1bcb6-23cd-41a4-a98d-ce5360928b43" containerID="291503256a24c9361b94a48bb468f7a22eb93773e030878cdf3221d52e7e7adc" exitCode=0 Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.595734 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wwcp6" event={"ID":"9af1bcb6-23cd-41a4-a98d-ce5360928b43","Type":"ContainerDied","Data":"291503256a24c9361b94a48bb468f7a22eb93773e030878cdf3221d52e7e7adc"} Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.595775 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wwcp6" event={"ID":"9af1bcb6-23cd-41a4-a98d-ce5360928b43","Type":"ContainerDied","Data":"fe23e9c4350dddf780ac2ac9cead8e06f1565029f24bcb1298bcc72c387a9ce7"} Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.595798 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wwcp6" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.595812 4868 scope.go:117] "RemoveContainer" containerID="291503256a24c9361b94a48bb468f7a22eb93773e030878cdf3221d52e7e7adc" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.636410 4868 scope.go:117] "RemoveContainer" containerID="5f0d6e123c3344427c4eb5c194d73105d629469ad9427f15e116c0eb0452fec1" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.642228 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wwcp6"] Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.651814 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wwcp6"] Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.659275 4868 scope.go:117] "RemoveContainer" containerID="e3c75bd92e00573f144fa2c1ba053a044d5d9cba77a81b3dc14b138cc96093ba" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.710131 4868 scope.go:117] "RemoveContainer" containerID="291503256a24c9361b94a48bb468f7a22eb93773e030878cdf3221d52e7e7adc" Dec 01 18:17:32 crc kubenswrapper[4868]: E1201 18:17:32.710678 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"291503256a24c9361b94a48bb468f7a22eb93773e030878cdf3221d52e7e7adc\": container with ID starting with 291503256a24c9361b94a48bb468f7a22eb93773e030878cdf3221d52e7e7adc not found: ID does not exist" containerID="291503256a24c9361b94a48bb468f7a22eb93773e030878cdf3221d52e7e7adc" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.710728 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"291503256a24c9361b94a48bb468f7a22eb93773e030878cdf3221d52e7e7adc"} err="failed to get container status \"291503256a24c9361b94a48bb468f7a22eb93773e030878cdf3221d52e7e7adc\": rpc error: code = NotFound desc = could not find container \"291503256a24c9361b94a48bb468f7a22eb93773e030878cdf3221d52e7e7adc\": container with ID starting with 291503256a24c9361b94a48bb468f7a22eb93773e030878cdf3221d52e7e7adc not found: ID does not exist" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.710756 4868 scope.go:117] "RemoveContainer" containerID="5f0d6e123c3344427c4eb5c194d73105d629469ad9427f15e116c0eb0452fec1" Dec 01 18:17:32 crc kubenswrapper[4868]: E1201 18:17:32.711141 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f0d6e123c3344427c4eb5c194d73105d629469ad9427f15e116c0eb0452fec1\": container with ID starting with 5f0d6e123c3344427c4eb5c194d73105d629469ad9427f15e116c0eb0452fec1 not found: ID does not exist" containerID="5f0d6e123c3344427c4eb5c194d73105d629469ad9427f15e116c0eb0452fec1" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.711162 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f0d6e123c3344427c4eb5c194d73105d629469ad9427f15e116c0eb0452fec1"} err="failed to get container status \"5f0d6e123c3344427c4eb5c194d73105d629469ad9427f15e116c0eb0452fec1\": rpc error: code = NotFound desc = could not find container \"5f0d6e123c3344427c4eb5c194d73105d629469ad9427f15e116c0eb0452fec1\": container with ID starting with 5f0d6e123c3344427c4eb5c194d73105d629469ad9427f15e116c0eb0452fec1 not found: ID does not exist" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.711178 4868 scope.go:117] "RemoveContainer" containerID="e3c75bd92e00573f144fa2c1ba053a044d5d9cba77a81b3dc14b138cc96093ba" Dec 01 18:17:32 crc kubenswrapper[4868]: E1201 18:17:32.711410 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3c75bd92e00573f144fa2c1ba053a044d5d9cba77a81b3dc14b138cc96093ba\": container with ID starting with e3c75bd92e00573f144fa2c1ba053a044d5d9cba77a81b3dc14b138cc96093ba not found: ID does not exist" containerID="e3c75bd92e00573f144fa2c1ba053a044d5d9cba77a81b3dc14b138cc96093ba" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.711431 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3c75bd92e00573f144fa2c1ba053a044d5d9cba77a81b3dc14b138cc96093ba"} err="failed to get container status \"e3c75bd92e00573f144fa2c1ba053a044d5d9cba77a81b3dc14b138cc96093ba\": rpc error: code = NotFound desc = could not find container \"e3c75bd92e00573f144fa2c1ba053a044d5d9cba77a81b3dc14b138cc96093ba\": container with ID starting with e3c75bd92e00573f144fa2c1ba053a044d5d9cba77a81b3dc14b138cc96093ba not found: ID does not exist" Dec 01 18:17:32 crc kubenswrapper[4868]: I1201 18:17:32.991668 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:33 crc kubenswrapper[4868]: I1201 18:17:33.043425 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:34 crc kubenswrapper[4868]: I1201 18:17:34.184884 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9af1bcb6-23cd-41a4-a98d-ce5360928b43" path="/var/lib/kubelet/pods/9af1bcb6-23cd-41a4-a98d-ce5360928b43/volumes" Dec 01 18:17:34 crc kubenswrapper[4868]: I1201 18:17:34.473049 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8trm7"] Dec 01 18:17:34 crc kubenswrapper[4868]: I1201 18:17:34.618537 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8trm7" podUID="30c1bdc3-084e-4f96-874b-95c1b143b485" containerName="registry-server" containerID="cri-o://b9b999de370b6d397a37f23c6a0234a545fc002141e10c5ec8f28e7f614a234a" gracePeriod=2 Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.111811 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.183263 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xrh4\" (UniqueName: \"kubernetes.io/projected/30c1bdc3-084e-4f96-874b-95c1b143b485-kube-api-access-5xrh4\") pod \"30c1bdc3-084e-4f96-874b-95c1b143b485\" (UID: \"30c1bdc3-084e-4f96-874b-95c1b143b485\") " Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.183335 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c1bdc3-084e-4f96-874b-95c1b143b485-utilities\") pod \"30c1bdc3-084e-4f96-874b-95c1b143b485\" (UID: \"30c1bdc3-084e-4f96-874b-95c1b143b485\") " Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.183437 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c1bdc3-084e-4f96-874b-95c1b143b485-catalog-content\") pod \"30c1bdc3-084e-4f96-874b-95c1b143b485\" (UID: \"30c1bdc3-084e-4f96-874b-95c1b143b485\") " Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.184308 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30c1bdc3-084e-4f96-874b-95c1b143b485-utilities" (OuterVolumeSpecName: "utilities") pod "30c1bdc3-084e-4f96-874b-95c1b143b485" (UID: "30c1bdc3-084e-4f96-874b-95c1b143b485"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.188295 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30c1bdc3-084e-4f96-874b-95c1b143b485-kube-api-access-5xrh4" (OuterVolumeSpecName: "kube-api-access-5xrh4") pod "30c1bdc3-084e-4f96-874b-95c1b143b485" (UID: "30c1bdc3-084e-4f96-874b-95c1b143b485"). InnerVolumeSpecName "kube-api-access-5xrh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.286268 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xrh4\" (UniqueName: \"kubernetes.io/projected/30c1bdc3-084e-4f96-874b-95c1b143b485-kube-api-access-5xrh4\") on node \"crc\" DevicePath \"\"" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.286308 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c1bdc3-084e-4f96-874b-95c1b143b485-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.292522 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30c1bdc3-084e-4f96-874b-95c1b143b485-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30c1bdc3-084e-4f96-874b-95c1b143b485" (UID: "30c1bdc3-084e-4f96-874b-95c1b143b485"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.388616 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c1bdc3-084e-4f96-874b-95c1b143b485-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.628146 4868 generic.go:334] "Generic (PLEG): container finished" podID="30c1bdc3-084e-4f96-874b-95c1b143b485" containerID="b9b999de370b6d397a37f23c6a0234a545fc002141e10c5ec8f28e7f614a234a" exitCode=0 Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.628187 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8trm7" event={"ID":"30c1bdc3-084e-4f96-874b-95c1b143b485","Type":"ContainerDied","Data":"b9b999de370b6d397a37f23c6a0234a545fc002141e10c5ec8f28e7f614a234a"} Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.628217 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8trm7" event={"ID":"30c1bdc3-084e-4f96-874b-95c1b143b485","Type":"ContainerDied","Data":"27a59d8ee12e750b1edb0d6ea3c52a12fcbedc3cc18cd9fb76bdff197da3b4df"} Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.628237 4868 scope.go:117] "RemoveContainer" containerID="b9b999de370b6d397a37f23c6a0234a545fc002141e10c5ec8f28e7f614a234a" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.628376 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8trm7" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.655744 4868 scope.go:117] "RemoveContainer" containerID="cb7a9c00e4697f9c8092a32bcc7446d43e1ffc498b8d9170182407934dd32723" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.667176 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8trm7"] Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.678637 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8trm7"] Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.696170 4868 scope.go:117] "RemoveContainer" containerID="f105bdabee44e7954ff513192f754af192ef7a2fe810236288ea54a66f2ded4d" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.734629 4868 scope.go:117] "RemoveContainer" containerID="b9b999de370b6d397a37f23c6a0234a545fc002141e10c5ec8f28e7f614a234a" Dec 01 18:17:35 crc kubenswrapper[4868]: E1201 18:17:35.735202 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9b999de370b6d397a37f23c6a0234a545fc002141e10c5ec8f28e7f614a234a\": container with ID starting with b9b999de370b6d397a37f23c6a0234a545fc002141e10c5ec8f28e7f614a234a not found: ID does not exist" containerID="b9b999de370b6d397a37f23c6a0234a545fc002141e10c5ec8f28e7f614a234a" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.735252 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9b999de370b6d397a37f23c6a0234a545fc002141e10c5ec8f28e7f614a234a"} err="failed to get container status \"b9b999de370b6d397a37f23c6a0234a545fc002141e10c5ec8f28e7f614a234a\": rpc error: code = NotFound desc = could not find container \"b9b999de370b6d397a37f23c6a0234a545fc002141e10c5ec8f28e7f614a234a\": container with ID starting with b9b999de370b6d397a37f23c6a0234a545fc002141e10c5ec8f28e7f614a234a not found: ID does not exist" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.735286 4868 scope.go:117] "RemoveContainer" containerID="cb7a9c00e4697f9c8092a32bcc7446d43e1ffc498b8d9170182407934dd32723" Dec 01 18:17:35 crc kubenswrapper[4868]: E1201 18:17:35.735807 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb7a9c00e4697f9c8092a32bcc7446d43e1ffc498b8d9170182407934dd32723\": container with ID starting with cb7a9c00e4697f9c8092a32bcc7446d43e1ffc498b8d9170182407934dd32723 not found: ID does not exist" containerID="cb7a9c00e4697f9c8092a32bcc7446d43e1ffc498b8d9170182407934dd32723" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.735856 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb7a9c00e4697f9c8092a32bcc7446d43e1ffc498b8d9170182407934dd32723"} err="failed to get container status \"cb7a9c00e4697f9c8092a32bcc7446d43e1ffc498b8d9170182407934dd32723\": rpc error: code = NotFound desc = could not find container \"cb7a9c00e4697f9c8092a32bcc7446d43e1ffc498b8d9170182407934dd32723\": container with ID starting with cb7a9c00e4697f9c8092a32bcc7446d43e1ffc498b8d9170182407934dd32723 not found: ID does not exist" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.735919 4868 scope.go:117] "RemoveContainer" containerID="f105bdabee44e7954ff513192f754af192ef7a2fe810236288ea54a66f2ded4d" Dec 01 18:17:35 crc kubenswrapper[4868]: E1201 18:17:35.736315 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f105bdabee44e7954ff513192f754af192ef7a2fe810236288ea54a66f2ded4d\": container with ID starting with f105bdabee44e7954ff513192f754af192ef7a2fe810236288ea54a66f2ded4d not found: ID does not exist" containerID="f105bdabee44e7954ff513192f754af192ef7a2fe810236288ea54a66f2ded4d" Dec 01 18:17:35 crc kubenswrapper[4868]: I1201 18:17:35.736345 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f105bdabee44e7954ff513192f754af192ef7a2fe810236288ea54a66f2ded4d"} err="failed to get container status \"f105bdabee44e7954ff513192f754af192ef7a2fe810236288ea54a66f2ded4d\": rpc error: code = NotFound desc = could not find container \"f105bdabee44e7954ff513192f754af192ef7a2fe810236288ea54a66f2ded4d\": container with ID starting with f105bdabee44e7954ff513192f754af192ef7a2fe810236288ea54a66f2ded4d not found: ID does not exist" Dec 01 18:17:36 crc kubenswrapper[4868]: I1201 18:17:36.183704 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30c1bdc3-084e-4f96-874b-95c1b143b485" path="/var/lib/kubelet/pods/30c1bdc3-084e-4f96-874b-95c1b143b485/volumes" Dec 01 18:17:40 crc kubenswrapper[4868]: I1201 18:17:40.172643 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:17:40 crc kubenswrapper[4868]: E1201 18:17:40.173371 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:17:53 crc kubenswrapper[4868]: I1201 18:17:53.171601 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:17:53 crc kubenswrapper[4868]: E1201 18:17:53.172331 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:18:06 crc kubenswrapper[4868]: I1201 18:18:06.172460 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:18:06 crc kubenswrapper[4868]: E1201 18:18:06.173294 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:18:17 crc kubenswrapper[4868]: I1201 18:18:17.173286 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:18:17 crc kubenswrapper[4868]: E1201 18:18:17.174384 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:18:31 crc kubenswrapper[4868]: I1201 18:18:31.171619 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:18:31 crc kubenswrapper[4868]: E1201 18:18:31.173227 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:18:45 crc kubenswrapper[4868]: I1201 18:18:45.172151 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:18:45 crc kubenswrapper[4868]: E1201 18:18:45.172930 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:19:00 crc kubenswrapper[4868]: I1201 18:19:00.172239 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:19:00 crc kubenswrapper[4868]: E1201 18:19:00.173074 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:19:04 crc kubenswrapper[4868]: I1201 18:19:04.498503 4868 generic.go:334] "Generic (PLEG): container finished" podID="8447d0d7-b9fd-4a48-bf54-e43272539aec" containerID="ad95d425eea60265bc1bc6b17537f4fba2176f26007dd3eff5ef8b09335af438" exitCode=0 Dec 01 18:19:04 crc kubenswrapper[4868]: I1201 18:19:04.498577 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8447d0d7-b9fd-4a48-bf54-e43272539aec","Type":"ContainerDied","Data":"ad95d425eea60265bc1bc6b17537f4fba2176f26007dd3eff5ef8b09335af438"} Dec 01 18:19:05 crc kubenswrapper[4868]: I1201 18:19:05.917835 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.069402 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"8447d0d7-b9fd-4a48-bf54-e43272539aec\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.069446 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-ssh-key\") pod \"8447d0d7-b9fd-4a48-bf54-e43272539aec\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.069502 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8447d0d7-b9fd-4a48-bf54-e43272539aec-openstack-config\") pod \"8447d0d7-b9fd-4a48-bf54-e43272539aec\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.069548 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8447d0d7-b9fd-4a48-bf54-e43272539aec-test-operator-ephemeral-workdir\") pod \"8447d0d7-b9fd-4a48-bf54-e43272539aec\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.069651 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8447d0d7-b9fd-4a48-bf54-e43272539aec-config-data\") pod \"8447d0d7-b9fd-4a48-bf54-e43272539aec\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.069674 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-openstack-config-secret\") pod \"8447d0d7-b9fd-4a48-bf54-e43272539aec\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.069697 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szcfg\" (UniqueName: \"kubernetes.io/projected/8447d0d7-b9fd-4a48-bf54-e43272539aec-kube-api-access-szcfg\") pod \"8447d0d7-b9fd-4a48-bf54-e43272539aec\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.069745 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-ca-certs\") pod \"8447d0d7-b9fd-4a48-bf54-e43272539aec\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.069774 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8447d0d7-b9fd-4a48-bf54-e43272539aec-test-operator-ephemeral-temporary\") pod \"8447d0d7-b9fd-4a48-bf54-e43272539aec\" (UID: \"8447d0d7-b9fd-4a48-bf54-e43272539aec\") " Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.070874 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8447d0d7-b9fd-4a48-bf54-e43272539aec-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "8447d0d7-b9fd-4a48-bf54-e43272539aec" (UID: "8447d0d7-b9fd-4a48-bf54-e43272539aec"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.071267 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8447d0d7-b9fd-4a48-bf54-e43272539aec-config-data" (OuterVolumeSpecName: "config-data") pod "8447d0d7-b9fd-4a48-bf54-e43272539aec" (UID: "8447d0d7-b9fd-4a48-bf54-e43272539aec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.075292 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8447d0d7-b9fd-4a48-bf54-e43272539aec-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "8447d0d7-b9fd-4a48-bf54-e43272539aec" (UID: "8447d0d7-b9fd-4a48-bf54-e43272539aec"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.077183 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "test-operator-logs") pod "8447d0d7-b9fd-4a48-bf54-e43272539aec" (UID: "8447d0d7-b9fd-4a48-bf54-e43272539aec"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.087228 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8447d0d7-b9fd-4a48-bf54-e43272539aec-kube-api-access-szcfg" (OuterVolumeSpecName: "kube-api-access-szcfg") pod "8447d0d7-b9fd-4a48-bf54-e43272539aec" (UID: "8447d0d7-b9fd-4a48-bf54-e43272539aec"). InnerVolumeSpecName "kube-api-access-szcfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.098580 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "8447d0d7-b9fd-4a48-bf54-e43272539aec" (UID: "8447d0d7-b9fd-4a48-bf54-e43272539aec"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.099771 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "8447d0d7-b9fd-4a48-bf54-e43272539aec" (UID: "8447d0d7-b9fd-4a48-bf54-e43272539aec"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.100587 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8447d0d7-b9fd-4a48-bf54-e43272539aec" (UID: "8447d0d7-b9fd-4a48-bf54-e43272539aec"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.131808 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8447d0d7-b9fd-4a48-bf54-e43272539aec-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "8447d0d7-b9fd-4a48-bf54-e43272539aec" (UID: "8447d0d7-b9fd-4a48-bf54-e43272539aec"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.172161 4868 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.172213 4868 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8447d0d7-b9fd-4a48-bf54-e43272539aec-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.172252 4868 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.172268 4868 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.172288 4868 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8447d0d7-b9fd-4a48-bf54-e43272539aec-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.172301 4868 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8447d0d7-b9fd-4a48-bf54-e43272539aec-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.172316 4868 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8447d0d7-b9fd-4a48-bf54-e43272539aec-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.172330 4868 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8447d0d7-b9fd-4a48-bf54-e43272539aec-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.172344 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szcfg\" (UniqueName: \"kubernetes.io/projected/8447d0d7-b9fd-4a48-bf54-e43272539aec-kube-api-access-szcfg\") on node \"crc\" DevicePath \"\"" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.194543 4868 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.273999 4868 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.517346 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8447d0d7-b9fd-4a48-bf54-e43272539aec","Type":"ContainerDied","Data":"695d6b3cf3d3bb90fb02b3e897b8e6f0c33e3b33663d5cb8dfd78c1c871a7104"} Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.517392 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="695d6b3cf3d3bb90fb02b3e897b8e6f0c33e3b33663d5cb8dfd78c1c871a7104" Dec 01 18:19:06 crc kubenswrapper[4868]: I1201 18:19:06.517523 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.425844 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 01 18:19:08 crc kubenswrapper[4868]: E1201 18:19:08.427862 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9af1bcb6-23cd-41a4-a98d-ce5360928b43" containerName="extract-utilities" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.427879 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="9af1bcb6-23cd-41a4-a98d-ce5360928b43" containerName="extract-utilities" Dec 01 18:19:08 crc kubenswrapper[4868]: E1201 18:19:08.427900 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60f2cfb2-9ecf-4666-b53b-681a6463e45b" containerName="extract-content" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.427907 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="60f2cfb2-9ecf-4666-b53b-681a6463e45b" containerName="extract-content" Dec 01 18:19:08 crc kubenswrapper[4868]: E1201 18:19:08.427925 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60f2cfb2-9ecf-4666-b53b-681a6463e45b" containerName="extract-utilities" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.427933 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="60f2cfb2-9ecf-4666-b53b-681a6463e45b" containerName="extract-utilities" Dec 01 18:19:08 crc kubenswrapper[4868]: E1201 18:19:08.427962 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9af1bcb6-23cd-41a4-a98d-ce5360928b43" containerName="extract-content" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.427969 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="9af1bcb6-23cd-41a4-a98d-ce5360928b43" containerName="extract-content" Dec 01 18:19:08 crc kubenswrapper[4868]: E1201 18:19:08.427986 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60f2cfb2-9ecf-4666-b53b-681a6463e45b" containerName="registry-server" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.427992 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="60f2cfb2-9ecf-4666-b53b-681a6463e45b" containerName="registry-server" Dec 01 18:19:08 crc kubenswrapper[4868]: E1201 18:19:08.428003 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8447d0d7-b9fd-4a48-bf54-e43272539aec" containerName="tempest-tests-tempest-tests-runner" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.428010 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="8447d0d7-b9fd-4a48-bf54-e43272539aec" containerName="tempest-tests-tempest-tests-runner" Dec 01 18:19:08 crc kubenswrapper[4868]: E1201 18:19:08.428020 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c1bdc3-084e-4f96-874b-95c1b143b485" containerName="extract-utilities" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.428026 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c1bdc3-084e-4f96-874b-95c1b143b485" containerName="extract-utilities" Dec 01 18:19:08 crc kubenswrapper[4868]: E1201 18:19:08.428044 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c1bdc3-084e-4f96-874b-95c1b143b485" containerName="registry-server" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.428050 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c1bdc3-084e-4f96-874b-95c1b143b485" containerName="registry-server" Dec 01 18:19:08 crc kubenswrapper[4868]: E1201 18:19:08.428064 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9af1bcb6-23cd-41a4-a98d-ce5360928b43" containerName="registry-server" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.428070 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="9af1bcb6-23cd-41a4-a98d-ce5360928b43" containerName="registry-server" Dec 01 18:19:08 crc kubenswrapper[4868]: E1201 18:19:08.428081 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c1bdc3-084e-4f96-874b-95c1b143b485" containerName="extract-content" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.428087 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c1bdc3-084e-4f96-874b-95c1b143b485" containerName="extract-content" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.428285 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="9af1bcb6-23cd-41a4-a98d-ce5360928b43" containerName="registry-server" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.428319 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="60f2cfb2-9ecf-4666-b53b-681a6463e45b" containerName="registry-server" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.428332 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="8447d0d7-b9fd-4a48-bf54-e43272539aec" containerName="tempest-tests-tempest-tests-runner" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.428346 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="30c1bdc3-084e-4f96-874b-95c1b143b485" containerName="registry-server" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.429244 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.431886 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vbmbb" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.438647 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.620709 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"66487491-7ab1-4ea0-b4df-e7be8a20a7e5\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.621168 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vsk2\" (UniqueName: \"kubernetes.io/projected/66487491-7ab1-4ea0-b4df-e7be8a20a7e5-kube-api-access-2vsk2\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"66487491-7ab1-4ea0-b4df-e7be8a20a7e5\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.723275 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"66487491-7ab1-4ea0-b4df-e7be8a20a7e5\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.723458 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vsk2\" (UniqueName: \"kubernetes.io/projected/66487491-7ab1-4ea0-b4df-e7be8a20a7e5-kube-api-access-2vsk2\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"66487491-7ab1-4ea0-b4df-e7be8a20a7e5\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.723704 4868 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"66487491-7ab1-4ea0-b4df-e7be8a20a7e5\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.748752 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vsk2\" (UniqueName: \"kubernetes.io/projected/66487491-7ab1-4ea0-b4df-e7be8a20a7e5-kube-api-access-2vsk2\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"66487491-7ab1-4ea0-b4df-e7be8a20a7e5\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 18:19:08 crc kubenswrapper[4868]: I1201 18:19:08.753227 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"66487491-7ab1-4ea0-b4df-e7be8a20a7e5\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 18:19:09 crc kubenswrapper[4868]: I1201 18:19:09.050230 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 18:19:09 crc kubenswrapper[4868]: I1201 18:19:09.553435 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 01 18:19:10 crc kubenswrapper[4868]: I1201 18:19:10.562783 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"66487491-7ab1-4ea0-b4df-e7be8a20a7e5","Type":"ContainerStarted","Data":"916383716ca30f192f23a107b669cde35290bc142537952460afb631b33b063c"} Dec 01 18:19:11 crc kubenswrapper[4868]: I1201 18:19:11.574628 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"66487491-7ab1-4ea0-b4df-e7be8a20a7e5","Type":"ContainerStarted","Data":"e8f6c1047a9dd7c15b9ed7fdd18dc9387f6804a9ad2ff0c950890c7c353c1423"} Dec 01 18:19:11 crc kubenswrapper[4868]: I1201 18:19:11.595529 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.659108686 podStartE2EDuration="3.595510337s" podCreationTimestamp="2025-12-01 18:19:08 +0000 UTC" firstStartedPulling="2025-12-01 18:19:09.567158884 +0000 UTC m=+3221.938269305" lastFinishedPulling="2025-12-01 18:19:10.503560545 +0000 UTC m=+3222.874670956" observedRunningTime="2025-12-01 18:19:11.588057285 +0000 UTC m=+3223.959167706" watchObservedRunningTime="2025-12-01 18:19:11.595510337 +0000 UTC m=+3223.966620758" Dec 01 18:19:15 crc kubenswrapper[4868]: I1201 18:19:15.172453 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:19:15 crc kubenswrapper[4868]: E1201 18:19:15.173516 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:19:26 crc kubenswrapper[4868]: I1201 18:19:26.171692 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:19:26 crc kubenswrapper[4868]: E1201 18:19:26.172785 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:19:33 crc kubenswrapper[4868]: I1201 18:19:33.374402 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5kwbg/must-gather-plcx8"] Dec 01 18:19:33 crc kubenswrapper[4868]: I1201 18:19:33.389532 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/must-gather-plcx8" Dec 01 18:19:33 crc kubenswrapper[4868]: I1201 18:19:33.392875 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-5kwbg"/"kube-root-ca.crt" Dec 01 18:19:33 crc kubenswrapper[4868]: I1201 18:19:33.398775 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-5kwbg"/"openshift-service-ca.crt" Dec 01 18:19:33 crc kubenswrapper[4868]: I1201 18:19:33.404085 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-5kwbg/must-gather-plcx8"] Dec 01 18:19:33 crc kubenswrapper[4868]: I1201 18:19:33.470213 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7z7g\" (UniqueName: \"kubernetes.io/projected/2a60cf14-31c5-461f-816f-4f97fd2999a5-kube-api-access-c7z7g\") pod \"must-gather-plcx8\" (UID: \"2a60cf14-31c5-461f-816f-4f97fd2999a5\") " pod="openshift-must-gather-5kwbg/must-gather-plcx8" Dec 01 18:19:33 crc kubenswrapper[4868]: I1201 18:19:33.470363 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2a60cf14-31c5-461f-816f-4f97fd2999a5-must-gather-output\") pod \"must-gather-plcx8\" (UID: \"2a60cf14-31c5-461f-816f-4f97fd2999a5\") " pod="openshift-must-gather-5kwbg/must-gather-plcx8" Dec 01 18:19:33 crc kubenswrapper[4868]: I1201 18:19:33.571890 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2a60cf14-31c5-461f-816f-4f97fd2999a5-must-gather-output\") pod \"must-gather-plcx8\" (UID: \"2a60cf14-31c5-461f-816f-4f97fd2999a5\") " pod="openshift-must-gather-5kwbg/must-gather-plcx8" Dec 01 18:19:33 crc kubenswrapper[4868]: I1201 18:19:33.571992 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7z7g\" (UniqueName: \"kubernetes.io/projected/2a60cf14-31c5-461f-816f-4f97fd2999a5-kube-api-access-c7z7g\") pod \"must-gather-plcx8\" (UID: \"2a60cf14-31c5-461f-816f-4f97fd2999a5\") " pod="openshift-must-gather-5kwbg/must-gather-plcx8" Dec 01 18:19:33 crc kubenswrapper[4868]: I1201 18:19:33.572511 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2a60cf14-31c5-461f-816f-4f97fd2999a5-must-gather-output\") pod \"must-gather-plcx8\" (UID: \"2a60cf14-31c5-461f-816f-4f97fd2999a5\") " pod="openshift-must-gather-5kwbg/must-gather-plcx8" Dec 01 18:19:33 crc kubenswrapper[4868]: I1201 18:19:33.591133 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7z7g\" (UniqueName: \"kubernetes.io/projected/2a60cf14-31c5-461f-816f-4f97fd2999a5-kube-api-access-c7z7g\") pod \"must-gather-plcx8\" (UID: \"2a60cf14-31c5-461f-816f-4f97fd2999a5\") " pod="openshift-must-gather-5kwbg/must-gather-plcx8" Dec 01 18:19:33 crc kubenswrapper[4868]: I1201 18:19:33.711767 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/must-gather-plcx8" Dec 01 18:19:34 crc kubenswrapper[4868]: I1201 18:19:34.190205 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-5kwbg/must-gather-plcx8"] Dec 01 18:19:34 crc kubenswrapper[4868]: W1201 18:19:34.192300 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a60cf14_31c5_461f_816f_4f97fd2999a5.slice/crio-70e7b9d69ecaf003c2f905b2d80dc51bd3531273a193295483b04541baa1c34f WatchSource:0}: Error finding container 70e7b9d69ecaf003c2f905b2d80dc51bd3531273a193295483b04541baa1c34f: Status 404 returned error can't find the container with id 70e7b9d69ecaf003c2f905b2d80dc51bd3531273a193295483b04541baa1c34f Dec 01 18:19:34 crc kubenswrapper[4868]: I1201 18:19:34.802414 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5kwbg/must-gather-plcx8" event={"ID":"2a60cf14-31c5-461f-816f-4f97fd2999a5","Type":"ContainerStarted","Data":"70e7b9d69ecaf003c2f905b2d80dc51bd3531273a193295483b04541baa1c34f"} Dec 01 18:19:38 crc kubenswrapper[4868]: I1201 18:19:38.837669 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5kwbg/must-gather-plcx8" event={"ID":"2a60cf14-31c5-461f-816f-4f97fd2999a5","Type":"ContainerStarted","Data":"07b747309324d285cf57258ad3a4c942bd9c38821e7bd8cb4a434eecf643f992"} Dec 01 18:19:38 crc kubenswrapper[4868]: I1201 18:19:38.838131 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5kwbg/must-gather-plcx8" event={"ID":"2a60cf14-31c5-461f-816f-4f97fd2999a5","Type":"ContainerStarted","Data":"417958dfa867323de1112a7435dd8a2354512c6bff43f60bcfdb99f869f9864b"} Dec 01 18:19:38 crc kubenswrapper[4868]: I1201 18:19:38.863673 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5kwbg/must-gather-plcx8" podStartSLOduration=2.108589342 podStartE2EDuration="5.863645789s" podCreationTimestamp="2025-12-01 18:19:33 +0000 UTC" firstStartedPulling="2025-12-01 18:19:34.195210898 +0000 UTC m=+3246.566321299" lastFinishedPulling="2025-12-01 18:19:37.950267335 +0000 UTC m=+3250.321377746" observedRunningTime="2025-12-01 18:19:38.854118531 +0000 UTC m=+3251.225228952" watchObservedRunningTime="2025-12-01 18:19:38.863645789 +0000 UTC m=+3251.234756200" Dec 01 18:19:40 crc kubenswrapper[4868]: I1201 18:19:40.172265 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:19:40 crc kubenswrapper[4868]: E1201 18:19:40.172903 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:19:40 crc kubenswrapper[4868]: E1201 18:19:40.832465 4868 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.103:47832->38.102.83.103:33157: write tcp 38.102.83.103:47832->38.102.83.103:33157: write: connection reset by peer Dec 01 18:19:41 crc kubenswrapper[4868]: I1201 18:19:41.913905 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5kwbg/crc-debug-lbbsn"] Dec 01 18:19:41 crc kubenswrapper[4868]: I1201 18:19:41.915408 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/crc-debug-lbbsn" Dec 01 18:19:41 crc kubenswrapper[4868]: I1201 18:19:41.917753 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-5kwbg"/"default-dockercfg-mjgbs" Dec 01 18:19:42 crc kubenswrapper[4868]: I1201 18:19:42.044197 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f33e503-9f88-44c3-9cc4-72ecc5763504-host\") pod \"crc-debug-lbbsn\" (UID: \"9f33e503-9f88-44c3-9cc4-72ecc5763504\") " pod="openshift-must-gather-5kwbg/crc-debug-lbbsn" Dec 01 18:19:42 crc kubenswrapper[4868]: I1201 18:19:42.044264 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgpnt\" (UniqueName: \"kubernetes.io/projected/9f33e503-9f88-44c3-9cc4-72ecc5763504-kube-api-access-zgpnt\") pod \"crc-debug-lbbsn\" (UID: \"9f33e503-9f88-44c3-9cc4-72ecc5763504\") " pod="openshift-must-gather-5kwbg/crc-debug-lbbsn" Dec 01 18:19:42 crc kubenswrapper[4868]: I1201 18:19:42.146241 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f33e503-9f88-44c3-9cc4-72ecc5763504-host\") pod \"crc-debug-lbbsn\" (UID: \"9f33e503-9f88-44c3-9cc4-72ecc5763504\") " pod="openshift-must-gather-5kwbg/crc-debug-lbbsn" Dec 01 18:19:42 crc kubenswrapper[4868]: I1201 18:19:42.146302 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgpnt\" (UniqueName: \"kubernetes.io/projected/9f33e503-9f88-44c3-9cc4-72ecc5763504-kube-api-access-zgpnt\") pod \"crc-debug-lbbsn\" (UID: \"9f33e503-9f88-44c3-9cc4-72ecc5763504\") " pod="openshift-must-gather-5kwbg/crc-debug-lbbsn" Dec 01 18:19:42 crc kubenswrapper[4868]: I1201 18:19:42.146423 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f33e503-9f88-44c3-9cc4-72ecc5763504-host\") pod \"crc-debug-lbbsn\" (UID: \"9f33e503-9f88-44c3-9cc4-72ecc5763504\") " pod="openshift-must-gather-5kwbg/crc-debug-lbbsn" Dec 01 18:19:42 crc kubenswrapper[4868]: I1201 18:19:42.169104 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgpnt\" (UniqueName: \"kubernetes.io/projected/9f33e503-9f88-44c3-9cc4-72ecc5763504-kube-api-access-zgpnt\") pod \"crc-debug-lbbsn\" (UID: \"9f33e503-9f88-44c3-9cc4-72ecc5763504\") " pod="openshift-must-gather-5kwbg/crc-debug-lbbsn" Dec 01 18:19:42 crc kubenswrapper[4868]: I1201 18:19:42.236429 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/crc-debug-lbbsn" Dec 01 18:19:42 crc kubenswrapper[4868]: W1201 18:19:42.290079 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f33e503_9f88_44c3_9cc4_72ecc5763504.slice/crio-b2d58b0b2cdb9f89068c19ede319c76e0839b1ace739b9b207e4c524de9535a7 WatchSource:0}: Error finding container b2d58b0b2cdb9f89068c19ede319c76e0839b1ace739b9b207e4c524de9535a7: Status 404 returned error can't find the container with id b2d58b0b2cdb9f89068c19ede319c76e0839b1ace739b9b207e4c524de9535a7 Dec 01 18:19:42 crc kubenswrapper[4868]: I1201 18:19:42.922759 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5kwbg/crc-debug-lbbsn" event={"ID":"9f33e503-9f88-44c3-9cc4-72ecc5763504","Type":"ContainerStarted","Data":"b2d58b0b2cdb9f89068c19ede319c76e0839b1ace739b9b207e4c524de9535a7"} Dec 01 18:19:53 crc kubenswrapper[4868]: I1201 18:19:53.173050 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:19:53 crc kubenswrapper[4868]: E1201 18:19:53.173811 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:19:55 crc kubenswrapper[4868]: I1201 18:19:55.066688 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5kwbg/crc-debug-lbbsn" event={"ID":"9f33e503-9f88-44c3-9cc4-72ecc5763504","Type":"ContainerStarted","Data":"bc4ed594d91529456ebc5cb17b988e45ef10d247116675aef7bd059e1c3b4378"} Dec 01 18:19:55 crc kubenswrapper[4868]: I1201 18:19:55.084021 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5kwbg/crc-debug-lbbsn" podStartSLOduration=2.22187827 podStartE2EDuration="14.083999219s" podCreationTimestamp="2025-12-01 18:19:41 +0000 UTC" firstStartedPulling="2025-12-01 18:19:42.292484531 +0000 UTC m=+3254.663594942" lastFinishedPulling="2025-12-01 18:19:54.15460548 +0000 UTC m=+3266.525715891" observedRunningTime="2025-12-01 18:19:55.081695937 +0000 UTC m=+3267.452806348" watchObservedRunningTime="2025-12-01 18:19:55.083999219 +0000 UTC m=+3267.455109630" Dec 01 18:20:05 crc kubenswrapper[4868]: I1201 18:20:05.173001 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:20:05 crc kubenswrapper[4868]: E1201 18:20:05.173743 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:20:20 crc kubenswrapper[4868]: I1201 18:20:20.173436 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:20:20 crc kubenswrapper[4868]: E1201 18:20:20.174077 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:20:35 crc kubenswrapper[4868]: I1201 18:20:35.171763 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:20:35 crc kubenswrapper[4868]: E1201 18:20:35.172496 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:20:39 crc kubenswrapper[4868]: I1201 18:20:39.792166 4868 generic.go:334] "Generic (PLEG): container finished" podID="9f33e503-9f88-44c3-9cc4-72ecc5763504" containerID="bc4ed594d91529456ebc5cb17b988e45ef10d247116675aef7bd059e1c3b4378" exitCode=0 Dec 01 18:20:39 crc kubenswrapper[4868]: I1201 18:20:39.792381 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5kwbg/crc-debug-lbbsn" event={"ID":"9f33e503-9f88-44c3-9cc4-72ecc5763504","Type":"ContainerDied","Data":"bc4ed594d91529456ebc5cb17b988e45ef10d247116675aef7bd059e1c3b4378"} Dec 01 18:20:40 crc kubenswrapper[4868]: I1201 18:20:40.926652 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/crc-debug-lbbsn" Dec 01 18:20:40 crc kubenswrapper[4868]: I1201 18:20:40.958340 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5kwbg/crc-debug-lbbsn"] Dec 01 18:20:40 crc kubenswrapper[4868]: I1201 18:20:40.966695 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5kwbg/crc-debug-lbbsn"] Dec 01 18:20:41 crc kubenswrapper[4868]: I1201 18:20:41.026623 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgpnt\" (UniqueName: \"kubernetes.io/projected/9f33e503-9f88-44c3-9cc4-72ecc5763504-kube-api-access-zgpnt\") pod \"9f33e503-9f88-44c3-9cc4-72ecc5763504\" (UID: \"9f33e503-9f88-44c3-9cc4-72ecc5763504\") " Dec 01 18:20:41 crc kubenswrapper[4868]: I1201 18:20:41.027084 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f33e503-9f88-44c3-9cc4-72ecc5763504-host\") pod \"9f33e503-9f88-44c3-9cc4-72ecc5763504\" (UID: \"9f33e503-9f88-44c3-9cc4-72ecc5763504\") " Dec 01 18:20:41 crc kubenswrapper[4868]: I1201 18:20:41.027153 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f33e503-9f88-44c3-9cc4-72ecc5763504-host" (OuterVolumeSpecName: "host") pod "9f33e503-9f88-44c3-9cc4-72ecc5763504" (UID: "9f33e503-9f88-44c3-9cc4-72ecc5763504"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 18:20:41 crc kubenswrapper[4868]: I1201 18:20:41.027592 4868 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f33e503-9f88-44c3-9cc4-72ecc5763504-host\") on node \"crc\" DevicePath \"\"" Dec 01 18:20:41 crc kubenswrapper[4868]: I1201 18:20:41.031706 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f33e503-9f88-44c3-9cc4-72ecc5763504-kube-api-access-zgpnt" (OuterVolumeSpecName: "kube-api-access-zgpnt") pod "9f33e503-9f88-44c3-9cc4-72ecc5763504" (UID: "9f33e503-9f88-44c3-9cc4-72ecc5763504"). InnerVolumeSpecName "kube-api-access-zgpnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:20:41 crc kubenswrapper[4868]: I1201 18:20:41.129450 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgpnt\" (UniqueName: \"kubernetes.io/projected/9f33e503-9f88-44c3-9cc4-72ecc5763504-kube-api-access-zgpnt\") on node \"crc\" DevicePath \"\"" Dec 01 18:20:41 crc kubenswrapper[4868]: I1201 18:20:41.812337 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2d58b0b2cdb9f89068c19ede319c76e0839b1ace739b9b207e4c524de9535a7" Dec 01 18:20:41 crc kubenswrapper[4868]: I1201 18:20:41.812671 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/crc-debug-lbbsn" Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.130372 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5kwbg/crc-debug-bxfds"] Dec 01 18:20:42 crc kubenswrapper[4868]: E1201 18:20:42.130867 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f33e503-9f88-44c3-9cc4-72ecc5763504" containerName="container-00" Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.130883 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f33e503-9f88-44c3-9cc4-72ecc5763504" containerName="container-00" Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.131152 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f33e503-9f88-44c3-9cc4-72ecc5763504" containerName="container-00" Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.131918 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/crc-debug-bxfds" Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.134161 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-5kwbg"/"default-dockercfg-mjgbs" Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.183810 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f33e503-9f88-44c3-9cc4-72ecc5763504" path="/var/lib/kubelet/pods/9f33e503-9f88-44c3-9cc4-72ecc5763504/volumes" Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.248445 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pf4k\" (UniqueName: \"kubernetes.io/projected/060cee3f-da0f-4f6c-b65f-83b970cfde1e-kube-api-access-7pf4k\") pod \"crc-debug-bxfds\" (UID: \"060cee3f-da0f-4f6c-b65f-83b970cfde1e\") " pod="openshift-must-gather-5kwbg/crc-debug-bxfds" Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.248535 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/060cee3f-da0f-4f6c-b65f-83b970cfde1e-host\") pod \"crc-debug-bxfds\" (UID: \"060cee3f-da0f-4f6c-b65f-83b970cfde1e\") " pod="openshift-must-gather-5kwbg/crc-debug-bxfds" Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.350315 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pf4k\" (UniqueName: \"kubernetes.io/projected/060cee3f-da0f-4f6c-b65f-83b970cfde1e-kube-api-access-7pf4k\") pod \"crc-debug-bxfds\" (UID: \"060cee3f-da0f-4f6c-b65f-83b970cfde1e\") " pod="openshift-must-gather-5kwbg/crc-debug-bxfds" Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.350423 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/060cee3f-da0f-4f6c-b65f-83b970cfde1e-host\") pod \"crc-debug-bxfds\" (UID: \"060cee3f-da0f-4f6c-b65f-83b970cfde1e\") " pod="openshift-must-gather-5kwbg/crc-debug-bxfds" Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.350643 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/060cee3f-da0f-4f6c-b65f-83b970cfde1e-host\") pod \"crc-debug-bxfds\" (UID: \"060cee3f-da0f-4f6c-b65f-83b970cfde1e\") " pod="openshift-must-gather-5kwbg/crc-debug-bxfds" Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.368795 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pf4k\" (UniqueName: \"kubernetes.io/projected/060cee3f-da0f-4f6c-b65f-83b970cfde1e-kube-api-access-7pf4k\") pod \"crc-debug-bxfds\" (UID: \"060cee3f-da0f-4f6c-b65f-83b970cfde1e\") " pod="openshift-must-gather-5kwbg/crc-debug-bxfds" Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.450867 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/crc-debug-bxfds" Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.824261 4868 generic.go:334] "Generic (PLEG): container finished" podID="060cee3f-da0f-4f6c-b65f-83b970cfde1e" containerID="c1419aecde3bfc7556ab97b0fbe5315e9d036fc734e1f46c1587d0981deab02e" exitCode=0 Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.824341 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5kwbg/crc-debug-bxfds" event={"ID":"060cee3f-da0f-4f6c-b65f-83b970cfde1e","Type":"ContainerDied","Data":"c1419aecde3bfc7556ab97b0fbe5315e9d036fc734e1f46c1587d0981deab02e"} Dec 01 18:20:42 crc kubenswrapper[4868]: I1201 18:20:42.824597 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5kwbg/crc-debug-bxfds" event={"ID":"060cee3f-da0f-4f6c-b65f-83b970cfde1e","Type":"ContainerStarted","Data":"d34d43587294312868fe2dcf02760554bcee252dc3f034fc0bd8b4b49bbae4b9"} Dec 01 18:20:43 crc kubenswrapper[4868]: I1201 18:20:43.345523 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5kwbg/crc-debug-bxfds"] Dec 01 18:20:43 crc kubenswrapper[4868]: I1201 18:20:43.355647 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5kwbg/crc-debug-bxfds"] Dec 01 18:20:43 crc kubenswrapper[4868]: I1201 18:20:43.931966 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/crc-debug-bxfds" Dec 01 18:20:43 crc kubenswrapper[4868]: I1201 18:20:43.988732 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pf4k\" (UniqueName: \"kubernetes.io/projected/060cee3f-da0f-4f6c-b65f-83b970cfde1e-kube-api-access-7pf4k\") pod \"060cee3f-da0f-4f6c-b65f-83b970cfde1e\" (UID: \"060cee3f-da0f-4f6c-b65f-83b970cfde1e\") " Dec 01 18:20:43 crc kubenswrapper[4868]: I1201 18:20:43.988876 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/060cee3f-da0f-4f6c-b65f-83b970cfde1e-host\") pod \"060cee3f-da0f-4f6c-b65f-83b970cfde1e\" (UID: \"060cee3f-da0f-4f6c-b65f-83b970cfde1e\") " Dec 01 18:20:43 crc kubenswrapper[4868]: I1201 18:20:43.989046 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/060cee3f-da0f-4f6c-b65f-83b970cfde1e-host" (OuterVolumeSpecName: "host") pod "060cee3f-da0f-4f6c-b65f-83b970cfde1e" (UID: "060cee3f-da0f-4f6c-b65f-83b970cfde1e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 18:20:43 crc kubenswrapper[4868]: I1201 18:20:43.989814 4868 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/060cee3f-da0f-4f6c-b65f-83b970cfde1e-host\") on node \"crc\" DevicePath \"\"" Dec 01 18:20:43 crc kubenswrapper[4868]: I1201 18:20:43.999060 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/060cee3f-da0f-4f6c-b65f-83b970cfde1e-kube-api-access-7pf4k" (OuterVolumeSpecName: "kube-api-access-7pf4k") pod "060cee3f-da0f-4f6c-b65f-83b970cfde1e" (UID: "060cee3f-da0f-4f6c-b65f-83b970cfde1e"). InnerVolumeSpecName "kube-api-access-7pf4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:20:44 crc kubenswrapper[4868]: I1201 18:20:44.091789 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pf4k\" (UniqueName: \"kubernetes.io/projected/060cee3f-da0f-4f6c-b65f-83b970cfde1e-kube-api-access-7pf4k\") on node \"crc\" DevicePath \"\"" Dec 01 18:20:44 crc kubenswrapper[4868]: I1201 18:20:44.187830 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="060cee3f-da0f-4f6c-b65f-83b970cfde1e" path="/var/lib/kubelet/pods/060cee3f-da0f-4f6c-b65f-83b970cfde1e/volumes" Dec 01 18:20:44 crc kubenswrapper[4868]: I1201 18:20:44.530404 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5kwbg/crc-debug-ctdnt"] Dec 01 18:20:44 crc kubenswrapper[4868]: E1201 18:20:44.531452 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="060cee3f-da0f-4f6c-b65f-83b970cfde1e" containerName="container-00" Dec 01 18:20:44 crc kubenswrapper[4868]: I1201 18:20:44.531473 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="060cee3f-da0f-4f6c-b65f-83b970cfde1e" containerName="container-00" Dec 01 18:20:44 crc kubenswrapper[4868]: I1201 18:20:44.531737 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="060cee3f-da0f-4f6c-b65f-83b970cfde1e" containerName="container-00" Dec 01 18:20:44 crc kubenswrapper[4868]: I1201 18:20:44.532730 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/crc-debug-ctdnt" Dec 01 18:20:44 crc kubenswrapper[4868]: I1201 18:20:44.599754 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ede4012a-922d-49da-b727-068ed96bf37c-host\") pod \"crc-debug-ctdnt\" (UID: \"ede4012a-922d-49da-b727-068ed96bf37c\") " pod="openshift-must-gather-5kwbg/crc-debug-ctdnt" Dec 01 18:20:44 crc kubenswrapper[4868]: I1201 18:20:44.599842 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr8vh\" (UniqueName: \"kubernetes.io/projected/ede4012a-922d-49da-b727-068ed96bf37c-kube-api-access-kr8vh\") pod \"crc-debug-ctdnt\" (UID: \"ede4012a-922d-49da-b727-068ed96bf37c\") " pod="openshift-must-gather-5kwbg/crc-debug-ctdnt" Dec 01 18:20:44 crc kubenswrapper[4868]: I1201 18:20:44.701534 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ede4012a-922d-49da-b727-068ed96bf37c-host\") pod \"crc-debug-ctdnt\" (UID: \"ede4012a-922d-49da-b727-068ed96bf37c\") " pod="openshift-must-gather-5kwbg/crc-debug-ctdnt" Dec 01 18:20:44 crc kubenswrapper[4868]: I1201 18:20:44.701614 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr8vh\" (UniqueName: \"kubernetes.io/projected/ede4012a-922d-49da-b727-068ed96bf37c-kube-api-access-kr8vh\") pod \"crc-debug-ctdnt\" (UID: \"ede4012a-922d-49da-b727-068ed96bf37c\") " pod="openshift-must-gather-5kwbg/crc-debug-ctdnt" Dec 01 18:20:44 crc kubenswrapper[4868]: I1201 18:20:44.701691 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ede4012a-922d-49da-b727-068ed96bf37c-host\") pod \"crc-debug-ctdnt\" (UID: \"ede4012a-922d-49da-b727-068ed96bf37c\") " pod="openshift-must-gather-5kwbg/crc-debug-ctdnt" Dec 01 18:20:44 crc kubenswrapper[4868]: I1201 18:20:44.726053 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr8vh\" (UniqueName: \"kubernetes.io/projected/ede4012a-922d-49da-b727-068ed96bf37c-kube-api-access-kr8vh\") pod \"crc-debug-ctdnt\" (UID: \"ede4012a-922d-49da-b727-068ed96bf37c\") " pod="openshift-must-gather-5kwbg/crc-debug-ctdnt" Dec 01 18:20:44 crc kubenswrapper[4868]: I1201 18:20:44.840667 4868 scope.go:117] "RemoveContainer" containerID="c1419aecde3bfc7556ab97b0fbe5315e9d036fc734e1f46c1587d0981deab02e" Dec 01 18:20:44 crc kubenswrapper[4868]: I1201 18:20:44.840702 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/crc-debug-bxfds" Dec 01 18:20:44 crc kubenswrapper[4868]: I1201 18:20:44.850623 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/crc-debug-ctdnt" Dec 01 18:20:44 crc kubenswrapper[4868]: W1201 18:20:44.897595 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podede4012a_922d_49da_b727_068ed96bf37c.slice/crio-3f40e56878f69acfd58d8dcfecd14997a27cfd66be97cbb5b93756e5f37deb58 WatchSource:0}: Error finding container 3f40e56878f69acfd58d8dcfecd14997a27cfd66be97cbb5b93756e5f37deb58: Status 404 returned error can't find the container with id 3f40e56878f69acfd58d8dcfecd14997a27cfd66be97cbb5b93756e5f37deb58 Dec 01 18:20:45 crc kubenswrapper[4868]: I1201 18:20:45.849862 4868 generic.go:334] "Generic (PLEG): container finished" podID="ede4012a-922d-49da-b727-068ed96bf37c" containerID="0601eba5b118ab023a450cc318371d85bf55bd366d258d6bf9ec3d64575c77a2" exitCode=0 Dec 01 18:20:45 crc kubenswrapper[4868]: I1201 18:20:45.849956 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5kwbg/crc-debug-ctdnt" event={"ID":"ede4012a-922d-49da-b727-068ed96bf37c","Type":"ContainerDied","Data":"0601eba5b118ab023a450cc318371d85bf55bd366d258d6bf9ec3d64575c77a2"} Dec 01 18:20:45 crc kubenswrapper[4868]: I1201 18:20:45.850365 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5kwbg/crc-debug-ctdnt" event={"ID":"ede4012a-922d-49da-b727-068ed96bf37c","Type":"ContainerStarted","Data":"3f40e56878f69acfd58d8dcfecd14997a27cfd66be97cbb5b93756e5f37deb58"} Dec 01 18:20:45 crc kubenswrapper[4868]: I1201 18:20:45.918193 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5kwbg/crc-debug-ctdnt"] Dec 01 18:20:45 crc kubenswrapper[4868]: I1201 18:20:45.934490 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5kwbg/crc-debug-ctdnt"] Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.580567 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s2rsv"] Dec 01 18:20:46 crc kubenswrapper[4868]: E1201 18:20:46.581248 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ede4012a-922d-49da-b727-068ed96bf37c" containerName="container-00" Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.581268 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="ede4012a-922d-49da-b727-068ed96bf37c" containerName="container-00" Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.581500 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="ede4012a-922d-49da-b727-068ed96bf37c" containerName="container-00" Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.582962 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.592370 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s2rsv"] Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.755541 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dbd7e01-7755-4477-a7ce-d976040ecd2f-catalog-content\") pod \"certified-operators-s2rsv\" (UID: \"5dbd7e01-7755-4477-a7ce-d976040ecd2f\") " pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.755597 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dbd7e01-7755-4477-a7ce-d976040ecd2f-utilities\") pod \"certified-operators-s2rsv\" (UID: \"5dbd7e01-7755-4477-a7ce-d976040ecd2f\") " pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.755840 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6bjw\" (UniqueName: \"kubernetes.io/projected/5dbd7e01-7755-4477-a7ce-d976040ecd2f-kube-api-access-s6bjw\") pod \"certified-operators-s2rsv\" (UID: \"5dbd7e01-7755-4477-a7ce-d976040ecd2f\") " pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.857274 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dbd7e01-7755-4477-a7ce-d976040ecd2f-catalog-content\") pod \"certified-operators-s2rsv\" (UID: \"5dbd7e01-7755-4477-a7ce-d976040ecd2f\") " pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.857326 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dbd7e01-7755-4477-a7ce-d976040ecd2f-utilities\") pod \"certified-operators-s2rsv\" (UID: \"5dbd7e01-7755-4477-a7ce-d976040ecd2f\") " pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.857395 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6bjw\" (UniqueName: \"kubernetes.io/projected/5dbd7e01-7755-4477-a7ce-d976040ecd2f-kube-api-access-s6bjw\") pod \"certified-operators-s2rsv\" (UID: \"5dbd7e01-7755-4477-a7ce-d976040ecd2f\") " pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.858182 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dbd7e01-7755-4477-a7ce-d976040ecd2f-catalog-content\") pod \"certified-operators-s2rsv\" (UID: \"5dbd7e01-7755-4477-a7ce-d976040ecd2f\") " pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.858207 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dbd7e01-7755-4477-a7ce-d976040ecd2f-utilities\") pod \"certified-operators-s2rsv\" (UID: \"5dbd7e01-7755-4477-a7ce-d976040ecd2f\") " pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.877820 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6bjw\" (UniqueName: \"kubernetes.io/projected/5dbd7e01-7755-4477-a7ce-d976040ecd2f-kube-api-access-s6bjw\") pod \"certified-operators-s2rsv\" (UID: \"5dbd7e01-7755-4477-a7ce-d976040ecd2f\") " pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.945956 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:46 crc kubenswrapper[4868]: I1201 18:20:46.974673 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/crc-debug-ctdnt" Dec 01 18:20:47 crc kubenswrapper[4868]: I1201 18:20:47.164480 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr8vh\" (UniqueName: \"kubernetes.io/projected/ede4012a-922d-49da-b727-068ed96bf37c-kube-api-access-kr8vh\") pod \"ede4012a-922d-49da-b727-068ed96bf37c\" (UID: \"ede4012a-922d-49da-b727-068ed96bf37c\") " Dec 01 18:20:47 crc kubenswrapper[4868]: I1201 18:20:47.164958 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ede4012a-922d-49da-b727-068ed96bf37c-host\") pod \"ede4012a-922d-49da-b727-068ed96bf37c\" (UID: \"ede4012a-922d-49da-b727-068ed96bf37c\") " Dec 01 18:20:47 crc kubenswrapper[4868]: I1201 18:20:47.165467 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ede4012a-922d-49da-b727-068ed96bf37c-host" (OuterVolumeSpecName: "host") pod "ede4012a-922d-49da-b727-068ed96bf37c" (UID: "ede4012a-922d-49da-b727-068ed96bf37c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 18:20:47 crc kubenswrapper[4868]: I1201 18:20:47.191258 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ede4012a-922d-49da-b727-068ed96bf37c-kube-api-access-kr8vh" (OuterVolumeSpecName: "kube-api-access-kr8vh") pod "ede4012a-922d-49da-b727-068ed96bf37c" (UID: "ede4012a-922d-49da-b727-068ed96bf37c"). InnerVolumeSpecName "kube-api-access-kr8vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:20:47 crc kubenswrapper[4868]: I1201 18:20:47.267919 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr8vh\" (UniqueName: \"kubernetes.io/projected/ede4012a-922d-49da-b727-068ed96bf37c-kube-api-access-kr8vh\") on node \"crc\" DevicePath \"\"" Dec 01 18:20:47 crc kubenswrapper[4868]: I1201 18:20:47.267977 4868 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ede4012a-922d-49da-b727-068ed96bf37c-host\") on node \"crc\" DevicePath \"\"" Dec 01 18:20:47 crc kubenswrapper[4868]: I1201 18:20:47.482738 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s2rsv"] Dec 01 18:20:47 crc kubenswrapper[4868]: W1201 18:20:47.486547 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5dbd7e01_7755_4477_a7ce_d976040ecd2f.slice/crio-d27b173525bd3859cdfe41d3958b018180cc9ae2c106906b98e76f7a157bd134 WatchSource:0}: Error finding container d27b173525bd3859cdfe41d3958b018180cc9ae2c106906b98e76f7a157bd134: Status 404 returned error can't find the container with id d27b173525bd3859cdfe41d3958b018180cc9ae2c106906b98e76f7a157bd134 Dec 01 18:20:47 crc kubenswrapper[4868]: I1201 18:20:47.875158 4868 generic.go:334] "Generic (PLEG): container finished" podID="5dbd7e01-7755-4477-a7ce-d976040ecd2f" containerID="c44f3ba1053e88cdf6b7a8006749cb537ad2ae4c4bae2175f4498270f3155062" exitCode=0 Dec 01 18:20:47 crc kubenswrapper[4868]: I1201 18:20:47.875218 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2rsv" event={"ID":"5dbd7e01-7755-4477-a7ce-d976040ecd2f","Type":"ContainerDied","Data":"c44f3ba1053e88cdf6b7a8006749cb537ad2ae4c4bae2175f4498270f3155062"} Dec 01 18:20:47 crc kubenswrapper[4868]: I1201 18:20:47.875512 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2rsv" event={"ID":"5dbd7e01-7755-4477-a7ce-d976040ecd2f","Type":"ContainerStarted","Data":"d27b173525bd3859cdfe41d3958b018180cc9ae2c106906b98e76f7a157bd134"} Dec 01 18:20:47 crc kubenswrapper[4868]: I1201 18:20:47.882633 4868 scope.go:117] "RemoveContainer" containerID="0601eba5b118ab023a450cc318371d85bf55bd366d258d6bf9ec3d64575c77a2" Dec 01 18:20:47 crc kubenswrapper[4868]: I1201 18:20:47.882710 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/crc-debug-ctdnt" Dec 01 18:20:48 crc kubenswrapper[4868]: I1201 18:20:48.186771 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ede4012a-922d-49da-b727-068ed96bf37c" path="/var/lib/kubelet/pods/ede4012a-922d-49da-b727-068ed96bf37c/volumes" Dec 01 18:20:49 crc kubenswrapper[4868]: I1201 18:20:49.907493 4868 generic.go:334] "Generic (PLEG): container finished" podID="5dbd7e01-7755-4477-a7ce-d976040ecd2f" containerID="1a845fc8116f9869276efe58802a04d65409fd0f59c3527c2c3496faeba2c4f7" exitCode=0 Dec 01 18:20:49 crc kubenswrapper[4868]: I1201 18:20:49.907671 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2rsv" event={"ID":"5dbd7e01-7755-4477-a7ce-d976040ecd2f","Type":"ContainerDied","Data":"1a845fc8116f9869276efe58802a04d65409fd0f59c3527c2c3496faeba2c4f7"} Dec 01 18:20:50 crc kubenswrapper[4868]: I1201 18:20:50.172692 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:20:50 crc kubenswrapper[4868]: E1201 18:20:50.173018 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:20:51 crc kubenswrapper[4868]: I1201 18:20:51.929260 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2rsv" event={"ID":"5dbd7e01-7755-4477-a7ce-d976040ecd2f","Type":"ContainerStarted","Data":"448d92b76f737db98821f531fe2a53accf48f946c74ffaa896be4bf66b1a27d0"} Dec 01 18:20:51 crc kubenswrapper[4868]: I1201 18:20:51.953864 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s2rsv" podStartSLOduration=3.075878045 podStartE2EDuration="5.953847483s" podCreationTimestamp="2025-12-01 18:20:46 +0000 UTC" firstStartedPulling="2025-12-01 18:20:47.878334075 +0000 UTC m=+3320.249444506" lastFinishedPulling="2025-12-01 18:20:50.756303533 +0000 UTC m=+3323.127413944" observedRunningTime="2025-12-01 18:20:51.948427505 +0000 UTC m=+3324.319537916" watchObservedRunningTime="2025-12-01 18:20:51.953847483 +0000 UTC m=+3324.324957884" Dec 01 18:20:56 crc kubenswrapper[4868]: I1201 18:20:56.946364 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:56 crc kubenswrapper[4868]: I1201 18:20:56.947037 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:56 crc kubenswrapper[4868]: I1201 18:20:56.995138 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:57 crc kubenswrapper[4868]: I1201 18:20:57.048381 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:57 crc kubenswrapper[4868]: I1201 18:20:57.232324 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s2rsv"] Dec 01 18:20:58 crc kubenswrapper[4868]: I1201 18:20:58.987658 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s2rsv" podUID="5dbd7e01-7755-4477-a7ce-d976040ecd2f" containerName="registry-server" containerID="cri-o://448d92b76f737db98821f531fe2a53accf48f946c74ffaa896be4bf66b1a27d0" gracePeriod=2 Dec 01 18:20:59 crc kubenswrapper[4868]: I1201 18:20:59.463888 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:20:59 crc kubenswrapper[4868]: I1201 18:20:59.500692 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dbd7e01-7755-4477-a7ce-d976040ecd2f-utilities\") pod \"5dbd7e01-7755-4477-a7ce-d976040ecd2f\" (UID: \"5dbd7e01-7755-4477-a7ce-d976040ecd2f\") " Dec 01 18:20:59 crc kubenswrapper[4868]: I1201 18:20:59.500748 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dbd7e01-7755-4477-a7ce-d976040ecd2f-catalog-content\") pod \"5dbd7e01-7755-4477-a7ce-d976040ecd2f\" (UID: \"5dbd7e01-7755-4477-a7ce-d976040ecd2f\") " Dec 01 18:20:59 crc kubenswrapper[4868]: I1201 18:20:59.501756 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dbd7e01-7755-4477-a7ce-d976040ecd2f-utilities" (OuterVolumeSpecName: "utilities") pod "5dbd7e01-7755-4477-a7ce-d976040ecd2f" (UID: "5dbd7e01-7755-4477-a7ce-d976040ecd2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:20:59 crc kubenswrapper[4868]: I1201 18:20:59.546126 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dbd7e01-7755-4477-a7ce-d976040ecd2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5dbd7e01-7755-4477-a7ce-d976040ecd2f" (UID: "5dbd7e01-7755-4477-a7ce-d976040ecd2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:20:59 crc kubenswrapper[4868]: I1201 18:20:59.602468 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6bjw\" (UniqueName: \"kubernetes.io/projected/5dbd7e01-7755-4477-a7ce-d976040ecd2f-kube-api-access-s6bjw\") pod \"5dbd7e01-7755-4477-a7ce-d976040ecd2f\" (UID: \"5dbd7e01-7755-4477-a7ce-d976040ecd2f\") " Dec 01 18:20:59 crc kubenswrapper[4868]: I1201 18:20:59.603651 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dbd7e01-7755-4477-a7ce-d976040ecd2f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 18:20:59 crc kubenswrapper[4868]: I1201 18:20:59.603674 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dbd7e01-7755-4477-a7ce-d976040ecd2f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 18:20:59 crc kubenswrapper[4868]: I1201 18:20:59.613774 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dbd7e01-7755-4477-a7ce-d976040ecd2f-kube-api-access-s6bjw" (OuterVolumeSpecName: "kube-api-access-s6bjw") pod "5dbd7e01-7755-4477-a7ce-d976040ecd2f" (UID: "5dbd7e01-7755-4477-a7ce-d976040ecd2f"). InnerVolumeSpecName "kube-api-access-s6bjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:20:59 crc kubenswrapper[4868]: I1201 18:20:59.705642 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6bjw\" (UniqueName: \"kubernetes.io/projected/5dbd7e01-7755-4477-a7ce-d976040ecd2f-kube-api-access-s6bjw\") on node \"crc\" DevicePath \"\"" Dec 01 18:20:59 crc kubenswrapper[4868]: I1201 18:20:59.997605 4868 generic.go:334] "Generic (PLEG): container finished" podID="5dbd7e01-7755-4477-a7ce-d976040ecd2f" containerID="448d92b76f737db98821f531fe2a53accf48f946c74ffaa896be4bf66b1a27d0" exitCode=0 Dec 01 18:20:59 crc kubenswrapper[4868]: I1201 18:20:59.997648 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2rsv" event={"ID":"5dbd7e01-7755-4477-a7ce-d976040ecd2f","Type":"ContainerDied","Data":"448d92b76f737db98821f531fe2a53accf48f946c74ffaa896be4bf66b1a27d0"} Dec 01 18:20:59 crc kubenswrapper[4868]: I1201 18:20:59.997675 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2rsv" event={"ID":"5dbd7e01-7755-4477-a7ce-d976040ecd2f","Type":"ContainerDied","Data":"d27b173525bd3859cdfe41d3958b018180cc9ae2c106906b98e76f7a157bd134"} Dec 01 18:20:59 crc kubenswrapper[4868]: I1201 18:20:59.997692 4868 scope.go:117] "RemoveContainer" containerID="448d92b76f737db98821f531fe2a53accf48f946c74ffaa896be4bf66b1a27d0" Dec 01 18:20:59 crc kubenswrapper[4868]: I1201 18:20:59.997802 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s2rsv" Dec 01 18:21:00 crc kubenswrapper[4868]: I1201 18:21:00.039808 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s2rsv"] Dec 01 18:21:00 crc kubenswrapper[4868]: I1201 18:21:00.046193 4868 scope.go:117] "RemoveContainer" containerID="1a845fc8116f9869276efe58802a04d65409fd0f59c3527c2c3496faeba2c4f7" Dec 01 18:21:00 crc kubenswrapper[4868]: I1201 18:21:00.049032 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s2rsv"] Dec 01 18:21:00 crc kubenswrapper[4868]: I1201 18:21:00.081132 4868 scope.go:117] "RemoveContainer" containerID="c44f3ba1053e88cdf6b7a8006749cb537ad2ae4c4bae2175f4498270f3155062" Dec 01 18:21:00 crc kubenswrapper[4868]: I1201 18:21:00.111698 4868 scope.go:117] "RemoveContainer" containerID="448d92b76f737db98821f531fe2a53accf48f946c74ffaa896be4bf66b1a27d0" Dec 01 18:21:00 crc kubenswrapper[4868]: E1201 18:21:00.112237 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"448d92b76f737db98821f531fe2a53accf48f946c74ffaa896be4bf66b1a27d0\": container with ID starting with 448d92b76f737db98821f531fe2a53accf48f946c74ffaa896be4bf66b1a27d0 not found: ID does not exist" containerID="448d92b76f737db98821f531fe2a53accf48f946c74ffaa896be4bf66b1a27d0" Dec 01 18:21:00 crc kubenswrapper[4868]: I1201 18:21:00.112285 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"448d92b76f737db98821f531fe2a53accf48f946c74ffaa896be4bf66b1a27d0"} err="failed to get container status \"448d92b76f737db98821f531fe2a53accf48f946c74ffaa896be4bf66b1a27d0\": rpc error: code = NotFound desc = could not find container \"448d92b76f737db98821f531fe2a53accf48f946c74ffaa896be4bf66b1a27d0\": container with ID starting with 448d92b76f737db98821f531fe2a53accf48f946c74ffaa896be4bf66b1a27d0 not found: ID does not exist" Dec 01 18:21:00 crc kubenswrapper[4868]: I1201 18:21:00.112316 4868 scope.go:117] "RemoveContainer" containerID="1a845fc8116f9869276efe58802a04d65409fd0f59c3527c2c3496faeba2c4f7" Dec 01 18:21:00 crc kubenswrapper[4868]: E1201 18:21:00.112777 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a845fc8116f9869276efe58802a04d65409fd0f59c3527c2c3496faeba2c4f7\": container with ID starting with 1a845fc8116f9869276efe58802a04d65409fd0f59c3527c2c3496faeba2c4f7 not found: ID does not exist" containerID="1a845fc8116f9869276efe58802a04d65409fd0f59c3527c2c3496faeba2c4f7" Dec 01 18:21:00 crc kubenswrapper[4868]: I1201 18:21:00.112812 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a845fc8116f9869276efe58802a04d65409fd0f59c3527c2c3496faeba2c4f7"} err="failed to get container status \"1a845fc8116f9869276efe58802a04d65409fd0f59c3527c2c3496faeba2c4f7\": rpc error: code = NotFound desc = could not find container \"1a845fc8116f9869276efe58802a04d65409fd0f59c3527c2c3496faeba2c4f7\": container with ID starting with 1a845fc8116f9869276efe58802a04d65409fd0f59c3527c2c3496faeba2c4f7 not found: ID does not exist" Dec 01 18:21:00 crc kubenswrapper[4868]: I1201 18:21:00.112855 4868 scope.go:117] "RemoveContainer" containerID="c44f3ba1053e88cdf6b7a8006749cb537ad2ae4c4bae2175f4498270f3155062" Dec 01 18:21:00 crc kubenswrapper[4868]: E1201 18:21:00.113310 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c44f3ba1053e88cdf6b7a8006749cb537ad2ae4c4bae2175f4498270f3155062\": container with ID starting with c44f3ba1053e88cdf6b7a8006749cb537ad2ae4c4bae2175f4498270f3155062 not found: ID does not exist" containerID="c44f3ba1053e88cdf6b7a8006749cb537ad2ae4c4bae2175f4498270f3155062" Dec 01 18:21:00 crc kubenswrapper[4868]: I1201 18:21:00.113352 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c44f3ba1053e88cdf6b7a8006749cb537ad2ae4c4bae2175f4498270f3155062"} err="failed to get container status \"c44f3ba1053e88cdf6b7a8006749cb537ad2ae4c4bae2175f4498270f3155062\": rpc error: code = NotFound desc = could not find container \"c44f3ba1053e88cdf6b7a8006749cb537ad2ae4c4bae2175f4498270f3155062\": container with ID starting with c44f3ba1053e88cdf6b7a8006749cb537ad2ae4c4bae2175f4498270f3155062 not found: ID does not exist" Dec 01 18:21:00 crc kubenswrapper[4868]: I1201 18:21:00.185349 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dbd7e01-7755-4477-a7ce-d976040ecd2f" path="/var/lib/kubelet/pods/5dbd7e01-7755-4477-a7ce-d976040ecd2f/volumes" Dec 01 18:21:02 crc kubenswrapper[4868]: I1201 18:21:02.534590 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-77d99f96cd-fz4zl_610e9c9c-cab0-4843-a120-5dde40363fd2/barbican-api/0.log" Dec 01 18:21:02 crc kubenswrapper[4868]: I1201 18:21:02.772457 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-77d99f96cd-fz4zl_610e9c9c-cab0-4843-a120-5dde40363fd2/barbican-api-log/0.log" Dec 01 18:21:02 crc kubenswrapper[4868]: I1201 18:21:02.832676 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79b8fddb46-xqzzs_0a4cab6a-d367-4eff-aab1-5b20b99f855e/barbican-keystone-listener/0.log" Dec 01 18:21:02 crc kubenswrapper[4868]: I1201 18:21:02.848282 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79b8fddb46-xqzzs_0a4cab6a-d367-4eff-aab1-5b20b99f855e/barbican-keystone-listener-log/0.log" Dec 01 18:21:03 crc kubenswrapper[4868]: I1201 18:21:03.016482 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-655b68fbf5-kxnvg_6fdfb743-319d-4a56-8182-442da79d3ed2/barbican-worker/0.log" Dec 01 18:21:03 crc kubenswrapper[4868]: I1201 18:21:03.103741 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-655b68fbf5-kxnvg_6fdfb743-319d-4a56-8182-442da79d3ed2/barbican-worker-log/0.log" Dec 01 18:21:03 crc kubenswrapper[4868]: I1201 18:21:03.172408 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:21:03 crc kubenswrapper[4868]: E1201 18:21:03.173176 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:21:03 crc kubenswrapper[4868]: I1201 18:21:03.231093 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp_af68e45f-ae90-4a6b-a189-7a97b42d0b95/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:03 crc kubenswrapper[4868]: I1201 18:21:03.342767 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0831ed8f-c5a0-4273-a32d-4f338a325073/ceilometer-notification-agent/0.log" Dec 01 18:21:03 crc kubenswrapper[4868]: I1201 18:21:03.361547 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0831ed8f-c5a0-4273-a32d-4f338a325073/ceilometer-central-agent/0.log" Dec 01 18:21:03 crc kubenswrapper[4868]: I1201 18:21:03.439850 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0831ed8f-c5a0-4273-a32d-4f338a325073/proxy-httpd/0.log" Dec 01 18:21:03 crc kubenswrapper[4868]: I1201 18:21:03.530854 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0831ed8f-c5a0-4273-a32d-4f338a325073/sg-core/0.log" Dec 01 18:21:03 crc kubenswrapper[4868]: I1201 18:21:03.638905 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_76a7855d-b030-49b6-a58f-4cdfc45d489a/cinder-api/0.log" Dec 01 18:21:03 crc kubenswrapper[4868]: I1201 18:21:03.696702 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_76a7855d-b030-49b6-a58f-4cdfc45d489a/cinder-api-log/0.log" Dec 01 18:21:03 crc kubenswrapper[4868]: I1201 18:21:03.866514 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d8a69c3d-3924-4160-87ec-17a00572a260/cinder-scheduler/0.log" Dec 01 18:21:03 crc kubenswrapper[4868]: I1201 18:21:03.943342 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d8a69c3d-3924-4160-87ec-17a00572a260/probe/0.log" Dec 01 18:21:04 crc kubenswrapper[4868]: I1201 18:21:04.062817 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg_fb542a8c-69c4-4a30-9a43-de20584d8e23/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:04 crc kubenswrapper[4868]: I1201 18:21:04.204109 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8_0d17c1fe-ab49-4510-a2cc-ed17ae06abad/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:04 crc kubenswrapper[4868]: I1201 18:21:04.260688 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-4rnqg_7480b802-459b-4029-9bc3-e3d7ba5099de/init/0.log" Dec 01 18:21:04 crc kubenswrapper[4868]: I1201 18:21:04.462761 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-4rnqg_7480b802-459b-4029-9bc3-e3d7ba5099de/init/0.log" Dec 01 18:21:04 crc kubenswrapper[4868]: I1201 18:21:04.477295 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-4rnqg_7480b802-459b-4029-9bc3-e3d7ba5099de/dnsmasq-dns/0.log" Dec 01 18:21:04 crc kubenswrapper[4868]: I1201 18:21:04.528760 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b_6e39e56d-e163-4e75-909f-410c72fde805/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:04 crc kubenswrapper[4868]: I1201 18:21:04.698049 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4247f042-96f0-4bc5-9ca3-a64ee5616624/glance-httpd/0.log" Dec 01 18:21:04 crc kubenswrapper[4868]: I1201 18:21:04.773412 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4247f042-96f0-4bc5-9ca3-a64ee5616624/glance-log/0.log" Dec 01 18:21:04 crc kubenswrapper[4868]: I1201 18:21:04.937253 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_2bc0d7b8-d497-4339-ad24-c9b75e6e06fc/glance-httpd/0.log" Dec 01 18:21:04 crc kubenswrapper[4868]: I1201 18:21:04.938809 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_2bc0d7b8-d497-4339-ad24-c9b75e6e06fc/glance-log/0.log" Dec 01 18:21:05 crc kubenswrapper[4868]: I1201 18:21:05.230770 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-77567b6ccb-79skk_c50007b0-2158-4940-ac84-c60384469583/horizon/0.log" Dec 01 18:21:05 crc kubenswrapper[4868]: I1201 18:21:05.290339 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-drbph_31db6887-0dfe-44c1-8951-757f2287cd9e/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:05 crc kubenswrapper[4868]: I1201 18:21:05.487000 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-77567b6ccb-79skk_c50007b0-2158-4940-ac84-c60384469583/horizon-log/0.log" Dec 01 18:21:05 crc kubenswrapper[4868]: I1201 18:21:05.512338 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-5x8p9_23a40d7f-6024-42d5-9f5f-cbdbc2c8740c/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:05 crc kubenswrapper[4868]: I1201 18:21:05.777667 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29410201-xgqzm_42f06b5b-cb48-40da-8a82-f3af58170592/keystone-cron/0.log" Dec 01 18:21:05 crc kubenswrapper[4868]: I1201 18:21:05.888214 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5476cd6644-rgwc7_7366fab1-f630-4998-b189-d1b90a274235/keystone-api/0.log" Dec 01 18:21:05 crc kubenswrapper[4868]: I1201 18:21:05.962552 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_36fb76da-24cc-4d21-898e-107fd20eb6a0/kube-state-metrics/0.log" Dec 01 18:21:06 crc kubenswrapper[4868]: I1201 18:21:06.072708 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs_944aa8d9-9aa5-466a-82fd-9768a8b970cb/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:06 crc kubenswrapper[4868]: I1201 18:21:06.492194 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5c9f554f77-pjbhr_cb459e0b-9096-4bc1-b553-41156deb1e39/neutron-httpd/0.log" Dec 01 18:21:06 crc kubenswrapper[4868]: I1201 18:21:06.493831 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5c9f554f77-pjbhr_cb459e0b-9096-4bc1-b553-41156deb1e39/neutron-api/0.log" Dec 01 18:21:06 crc kubenswrapper[4868]: I1201 18:21:06.585418 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6_6201a05d-611d-4220-b7ae-1b0bf860b4c0/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:07 crc kubenswrapper[4868]: I1201 18:21:07.158355 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_47392f20-a05b-44d5-be29-1d7588516c0a/nova-api-log/0.log" Dec 01 18:21:07 crc kubenswrapper[4868]: I1201 18:21:07.208469 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_83a0dfb3-863c-4c80-b46c-e4e3b5fe4098/nova-cell0-conductor-conductor/0.log" Dec 01 18:21:07 crc kubenswrapper[4868]: I1201 18:21:07.370027 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_47392f20-a05b-44d5-be29-1d7588516c0a/nova-api-api/0.log" Dec 01 18:21:07 crc kubenswrapper[4868]: I1201 18:21:07.480876 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c/nova-cell1-conductor-conductor/0.log" Dec 01 18:21:07 crc kubenswrapper[4868]: I1201 18:21:07.512637 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_4014a0ba-e992-4592-aafd-62e2c90e2ab0/nova-cell1-novncproxy-novncproxy/0.log" Dec 01 18:21:07 crc kubenswrapper[4868]: I1201 18:21:07.721728 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-j6jdm_0739d3f5-f361-46f0-872d-23f9d29b7d46/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:07 crc kubenswrapper[4868]: I1201 18:21:07.878862 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_76578957-dead-4612-9d4d-bd9d19c589ef/nova-metadata-log/0.log" Dec 01 18:21:08 crc kubenswrapper[4868]: I1201 18:21:08.194844 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_cc48650c-5f3d-4c0a-b570-c7eb8615d504/mysql-bootstrap/0.log" Dec 01 18:21:08 crc kubenswrapper[4868]: I1201 18:21:08.197697 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_ea59e84e-774b-4d8f-af90-1de84fe578d7/nova-scheduler-scheduler/0.log" Dec 01 18:21:08 crc kubenswrapper[4868]: I1201 18:21:08.395711 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_cc48650c-5f3d-4c0a-b570-c7eb8615d504/galera/0.log" Dec 01 18:21:08 crc kubenswrapper[4868]: I1201 18:21:08.451261 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_cc48650c-5f3d-4c0a-b570-c7eb8615d504/mysql-bootstrap/0.log" Dec 01 18:21:08 crc kubenswrapper[4868]: I1201 18:21:08.596629 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_68f3c214-92ad-43c0-b06a-33c6f1f90cfb/mysql-bootstrap/0.log" Dec 01 18:21:08 crc kubenswrapper[4868]: I1201 18:21:08.810578 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_68f3c214-92ad-43c0-b06a-33c6f1f90cfb/mysql-bootstrap/0.log" Dec 01 18:21:08 crc kubenswrapper[4868]: I1201 18:21:08.855315 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_68f3c214-92ad-43c0-b06a-33c6f1f90cfb/galera/0.log" Dec 01 18:21:09 crc kubenswrapper[4868]: I1201 18:21:09.002663 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_76578957-dead-4612-9d4d-bd9d19c589ef/nova-metadata-metadata/0.log" Dec 01 18:21:09 crc kubenswrapper[4868]: I1201 18:21:09.024618 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_3f3f5b1c-3ed5-4d79-b298-474f483b2c4a/openstackclient/0.log" Dec 01 18:21:09 crc kubenswrapper[4868]: I1201 18:21:09.107614 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-87c94_7328948b-0a8c-45d3-91b1-c8317bc12499/openstack-network-exporter/0.log" Dec 01 18:21:09 crc kubenswrapper[4868]: I1201 18:21:09.254128 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rzgnl_8cca4bfb-cddc-425e-abe2-6debea4aa92a/ovsdb-server-init/0.log" Dec 01 18:21:09 crc kubenswrapper[4868]: I1201 18:21:09.418117 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rzgnl_8cca4bfb-cddc-425e-abe2-6debea4aa92a/ovs-vswitchd/0.log" Dec 01 18:21:09 crc kubenswrapper[4868]: I1201 18:21:09.428311 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rzgnl_8cca4bfb-cddc-425e-abe2-6debea4aa92a/ovsdb-server-init/0.log" Dec 01 18:21:09 crc kubenswrapper[4868]: I1201 18:21:09.482515 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rzgnl_8cca4bfb-cddc-425e-abe2-6debea4aa92a/ovsdb-server/0.log" Dec 01 18:21:09 crc kubenswrapper[4868]: I1201 18:21:09.604177 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-z4f4z_65d7a832-930e-4103-90f6-dbc5de8c1ece/ovn-controller/0.log" Dec 01 18:21:09 crc kubenswrapper[4868]: I1201 18:21:09.767169 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-nmn4w_216fe1b0-ec90-4ee7-91e4-aa24476b39b4/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:09 crc kubenswrapper[4868]: I1201 18:21:09.870284 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_56ae3649-0322-453b-9088-6807f59d7d96/openstack-network-exporter/0.log" Dec 01 18:21:09 crc kubenswrapper[4868]: I1201 18:21:09.962655 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_56ae3649-0322-453b-9088-6807f59d7d96/ovn-northd/0.log" Dec 01 18:21:10 crc kubenswrapper[4868]: I1201 18:21:10.084161 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_faa33685-51f4-46ce-bdc7-ef9a2467f482/openstack-network-exporter/0.log" Dec 01 18:21:10 crc kubenswrapper[4868]: I1201 18:21:10.137789 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_faa33685-51f4-46ce-bdc7-ef9a2467f482/ovsdbserver-nb/0.log" Dec 01 18:21:10 crc kubenswrapper[4868]: I1201 18:21:10.266857 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_89b376d8-f72b-4c03-a2f1-3fcf6ff20336/openstack-network-exporter/0.log" Dec 01 18:21:10 crc kubenswrapper[4868]: I1201 18:21:10.304993 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_89b376d8-f72b-4c03-a2f1-3fcf6ff20336/ovsdbserver-sb/0.log" Dec 01 18:21:10 crc kubenswrapper[4868]: I1201 18:21:10.514404 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d8466bd7b-sr5pl_f41cf3a5-eb10-42c0-93a1-f1195331f807/placement-api/0.log" Dec 01 18:21:10 crc kubenswrapper[4868]: I1201 18:21:10.561682 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d8466bd7b-sr5pl_f41cf3a5-eb10-42c0-93a1-f1195331f807/placement-log/0.log" Dec 01 18:21:10 crc kubenswrapper[4868]: I1201 18:21:10.594691 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_12c9fc00-91a0-493e-b6b3-5d1a5ba7685c/setup-container/0.log" Dec 01 18:21:10 crc kubenswrapper[4868]: I1201 18:21:10.845912 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_12c9fc00-91a0-493e-b6b3-5d1a5ba7685c/setup-container/0.log" Dec 01 18:21:10 crc kubenswrapper[4868]: I1201 18:21:10.933446 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b05cb910-a007-46d5-8182-eb1d30c5f953/setup-container/0.log" Dec 01 18:21:11 crc kubenswrapper[4868]: I1201 18:21:11.442813 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_12c9fc00-91a0-493e-b6b3-5d1a5ba7685c/rabbitmq/0.log" Dec 01 18:21:11 crc kubenswrapper[4868]: I1201 18:21:11.607536 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b05cb910-a007-46d5-8182-eb1d30c5f953/rabbitmq/0.log" Dec 01 18:21:11 crc kubenswrapper[4868]: I1201 18:21:11.621349 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b05cb910-a007-46d5-8182-eb1d30c5f953/setup-container/0.log" Dec 01 18:21:11 crc kubenswrapper[4868]: I1201 18:21:11.769726 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9_17fd50a0-00e4-4b51-9631-30b48bd3013f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:11 crc kubenswrapper[4868]: I1201 18:21:11.966916 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-vvbkm_e989defc-ac68-4417-9bf3-f5962bfdb601/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:12 crc kubenswrapper[4868]: I1201 18:21:12.015791 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb_f36695d5-3d20-4e3a-9ef3-63cc50a3445c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:12 crc kubenswrapper[4868]: I1201 18:21:12.219830 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-q9z9p_2163d223-acf7-452f-8a2e-5461307cf82e/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:12 crc kubenswrapper[4868]: I1201 18:21:12.396662 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-zjnp8_3f7f2561-e242-48b1-a363-50bdecab86de/ssh-known-hosts-edpm-deployment/0.log" Dec 01 18:21:12 crc kubenswrapper[4868]: I1201 18:21:12.554796 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-86984fb57-t5hzr_a427adaa-2060-4639-bd6d-e53b8fb00357/proxy-server/0.log" Dec 01 18:21:12 crc kubenswrapper[4868]: I1201 18:21:12.576291 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-86984fb57-t5hzr_a427adaa-2060-4639-bd6d-e53b8fb00357/proxy-httpd/0.log" Dec 01 18:21:12 crc kubenswrapper[4868]: I1201 18:21:12.703284 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-bfdvt_71e5e846-0b2a-47d3-b60b-daf23c22af52/swift-ring-rebalance/0.log" Dec 01 18:21:12 crc kubenswrapper[4868]: I1201 18:21:12.833128 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/account-auditor/0.log" Dec 01 18:21:13 crc kubenswrapper[4868]: I1201 18:21:13.139019 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/account-reaper/0.log" Dec 01 18:21:13 crc kubenswrapper[4868]: I1201 18:21:13.167193 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/account-replicator/0.log" Dec 01 18:21:13 crc kubenswrapper[4868]: I1201 18:21:13.302215 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/account-server/0.log" Dec 01 18:21:13 crc kubenswrapper[4868]: I1201 18:21:13.306084 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/container-auditor/0.log" Dec 01 18:21:13 crc kubenswrapper[4868]: I1201 18:21:13.371587 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/container-replicator/0.log" Dec 01 18:21:13 crc kubenswrapper[4868]: I1201 18:21:13.419625 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/container-server/0.log" Dec 01 18:21:13 crc kubenswrapper[4868]: I1201 18:21:13.530416 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/container-updater/0.log" Dec 01 18:21:13 crc kubenswrapper[4868]: I1201 18:21:13.542333 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/object-auditor/0.log" Dec 01 18:21:13 crc kubenswrapper[4868]: I1201 18:21:13.622271 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/object-expirer/0.log" Dec 01 18:21:13 crc kubenswrapper[4868]: I1201 18:21:13.704375 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/object-replicator/0.log" Dec 01 18:21:13 crc kubenswrapper[4868]: I1201 18:21:13.739061 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/object-server/0.log" Dec 01 18:21:13 crc kubenswrapper[4868]: I1201 18:21:13.747190 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/object-updater/0.log" Dec 01 18:21:13 crc kubenswrapper[4868]: I1201 18:21:13.903286 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/rsync/0.log" Dec 01 18:21:13 crc kubenswrapper[4868]: I1201 18:21:13.907994 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/swift-recon-cron/0.log" Dec 01 18:21:14 crc kubenswrapper[4868]: I1201 18:21:14.003175 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq_64998ba8-012c-4ffc-a7b3-dbbbb41940c0/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:14 crc kubenswrapper[4868]: I1201 18:21:14.149714 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_8447d0d7-b9fd-4a48-bf54-e43272539aec/tempest-tests-tempest-tests-runner/0.log" Dec 01 18:21:14 crc kubenswrapper[4868]: I1201 18:21:14.280106 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_66487491-7ab1-4ea0-b4df-e7be8a20a7e5/test-operator-logs-container/0.log" Dec 01 18:21:14 crc kubenswrapper[4868]: I1201 18:21:14.468903 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-v6t68_f1d81003-13ab-4833-9b01-3b4c6f8876ef/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:21:18 crc kubenswrapper[4868]: I1201 18:21:18.178525 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:21:18 crc kubenswrapper[4868]: E1201 18:21:18.181381 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:21:22 crc kubenswrapper[4868]: I1201 18:21:22.425456 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e/memcached/0.log" Dec 01 18:21:30 crc kubenswrapper[4868]: I1201 18:21:30.172226 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:21:30 crc kubenswrapper[4868]: E1201 18:21:30.173046 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:21:39 crc kubenswrapper[4868]: I1201 18:21:39.173592 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts_127f8203-d151-4430-97ae-405c8788a2af/util/0.log" Dec 01 18:21:39 crc kubenswrapper[4868]: I1201 18:21:39.410885 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts_127f8203-d151-4430-97ae-405c8788a2af/util/0.log" Dec 01 18:21:39 crc kubenswrapper[4868]: I1201 18:21:39.437641 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts_127f8203-d151-4430-97ae-405c8788a2af/pull/0.log" Dec 01 18:21:39 crc kubenswrapper[4868]: I1201 18:21:39.441787 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts_127f8203-d151-4430-97ae-405c8788a2af/pull/0.log" Dec 01 18:21:39 crc kubenswrapper[4868]: I1201 18:21:39.609675 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts_127f8203-d151-4430-97ae-405c8788a2af/pull/0.log" Dec 01 18:21:39 crc kubenswrapper[4868]: I1201 18:21:39.649526 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts_127f8203-d151-4430-97ae-405c8788a2af/util/0.log" Dec 01 18:21:39 crc kubenswrapper[4868]: I1201 18:21:39.699137 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts_127f8203-d151-4430-97ae-405c8788a2af/extract/0.log" Dec 01 18:21:39 crc kubenswrapper[4868]: I1201 18:21:39.839394 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-8m6s9_f3eeed9b-dfbf-4e17-b80e-8792d471f1fb/kube-rbac-proxy/0.log" Dec 01 18:21:39 crc kubenswrapper[4868]: I1201 18:21:39.947567 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-pk4dk_8cc7f247-f40c-4486-bd86-5f49036a4b2e/kube-rbac-proxy/0.log" Dec 01 18:21:39 crc kubenswrapper[4868]: I1201 18:21:39.969566 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-8m6s9_f3eeed9b-dfbf-4e17-b80e-8792d471f1fb/manager/0.log" Dec 01 18:21:40 crc kubenswrapper[4868]: I1201 18:21:40.090993 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-pk4dk_8cc7f247-f40c-4486-bd86-5f49036a4b2e/manager/0.log" Dec 01 18:21:40 crc kubenswrapper[4868]: I1201 18:21:40.154887 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-4s44v_a11d06ed-a89a-4600-8c63-4a3845eb01e6/kube-rbac-proxy/0.log" Dec 01 18:21:40 crc kubenswrapper[4868]: I1201 18:21:40.178269 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-4s44v_a11d06ed-a89a-4600-8c63-4a3845eb01e6/manager/0.log" Dec 01 18:21:40 crc kubenswrapper[4868]: I1201 18:21:40.291203 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-rhj2q_6faacf07-6354-4314-829b-db239c85d98e/kube-rbac-proxy/0.log" Dec 01 18:21:40 crc kubenswrapper[4868]: I1201 18:21:40.449880 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-rhj2q_6faacf07-6354-4314-829b-db239c85d98e/manager/0.log" Dec 01 18:21:40 crc kubenswrapper[4868]: I1201 18:21:40.472870 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-kbb5v_c5c836e6-8b19-44f8-91ae-4893dcbfd0fc/kube-rbac-proxy/0.log" Dec 01 18:21:40 crc kubenswrapper[4868]: I1201 18:21:40.570355 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-kbb5v_c5c836e6-8b19-44f8-91ae-4893dcbfd0fc/manager/0.log" Dec 01 18:21:40 crc kubenswrapper[4868]: I1201 18:21:40.655194 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5kkb2_93fc4a4e-c6fb-4d43-88dd-0c24bcd85c5e/kube-rbac-proxy/0.log" Dec 01 18:21:40 crc kubenswrapper[4868]: I1201 18:21:40.676120 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5kkb2_93fc4a4e-c6fb-4d43-88dd-0c24bcd85c5e/manager/0.log" Dec 01 18:21:41 crc kubenswrapper[4868]: I1201 18:21:41.032030 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-55599fd5c4-2vg29_775bad96-f2da-4628-8242-af4778e37aee/kube-rbac-proxy/0.log" Dec 01 18:21:41 crc kubenswrapper[4868]: I1201 18:21:41.144444 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-c9nqt_5cad124c-35fd-404e-bcf4-3b17e89e1288/kube-rbac-proxy/0.log" Dec 01 18:21:41 crc kubenswrapper[4868]: I1201 18:21:41.217474 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-55599fd5c4-2vg29_775bad96-f2da-4628-8242-af4778e37aee/manager/0.log" Dec 01 18:21:41 crc kubenswrapper[4868]: I1201 18:21:41.241773 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-c9nqt_5cad124c-35fd-404e-bcf4-3b17e89e1288/manager/0.log" Dec 01 18:21:41 crc kubenswrapper[4868]: I1201 18:21:41.398042 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-hqmbs_ccf55d21-7eac-4a3e-90c6-147287c4aac3/kube-rbac-proxy/0.log" Dec 01 18:21:41 crc kubenswrapper[4868]: I1201 18:21:41.482521 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-hqmbs_ccf55d21-7eac-4a3e-90c6-147287c4aac3/manager/0.log" Dec 01 18:21:41 crc kubenswrapper[4868]: I1201 18:21:41.613252 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-jjvvf_592806bc-2a42-4c69-b36f-1b8dc3c14249/kube-rbac-proxy/0.log" Dec 01 18:21:41 crc kubenswrapper[4868]: I1201 18:21:41.714203 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-jjvvf_592806bc-2a42-4c69-b36f-1b8dc3c14249/manager/0.log" Dec 01 18:21:41 crc kubenswrapper[4868]: I1201 18:21:41.800261 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-9j2nw_4e7cc027-32a6-437e-b05b-52c2984e0a61/kube-rbac-proxy/0.log" Dec 01 18:21:41 crc kubenswrapper[4868]: I1201 18:21:41.875024 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-9j2nw_4e7cc027-32a6-437e-b05b-52c2984e0a61/manager/0.log" Dec 01 18:21:41 crc kubenswrapper[4868]: I1201 18:21:41.907048 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-7mptb_97f09844-bdf4-4612-b4b4-966ae8dd49c2/kube-rbac-proxy/0.log" Dec 01 18:21:42 crc kubenswrapper[4868]: I1201 18:21:42.112829 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-7mptb_97f09844-bdf4-4612-b4b4-966ae8dd49c2/manager/0.log" Dec 01 18:21:42 crc kubenswrapper[4868]: I1201 18:21:42.158805 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-f7k69_b8df3e8b-0fc5-465f-b531-474518a1b809/kube-rbac-proxy/0.log" Dec 01 18:21:42 crc kubenswrapper[4868]: I1201 18:21:42.173253 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:21:42 crc kubenswrapper[4868]: E1201 18:21:42.173497 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:21:42 crc kubenswrapper[4868]: I1201 18:21:42.303440 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-f7k69_b8df3e8b-0fc5-465f-b531-474518a1b809/manager/0.log" Dec 01 18:21:42 crc kubenswrapper[4868]: I1201 18:21:42.373986 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-fmkmw_20c07d22-1628-40a4-97aa-605a2da611df/manager/0.log" Dec 01 18:21:42 crc kubenswrapper[4868]: I1201 18:21:42.410524 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-fmkmw_20c07d22-1628-40a4-97aa-605a2da611df/kube-rbac-proxy/0.log" Dec 01 18:21:42 crc kubenswrapper[4868]: I1201 18:21:42.537828 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs_25f93b71-f929-4f74-a2d9-822f0b9402bb/kube-rbac-proxy/0.log" Dec 01 18:21:42 crc kubenswrapper[4868]: I1201 18:21:42.662729 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs_25f93b71-f929-4f74-a2d9-822f0b9402bb/manager/0.log" Dec 01 18:21:43 crc kubenswrapper[4868]: I1201 18:21:43.108446 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-8rf2p_d73d228d-a5b1-4569-8c1f-a062812250b9/registry-server/0.log" Dec 01 18:21:43 crc kubenswrapper[4868]: I1201 18:21:43.123802 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7b559cfbd8-zw5pm_b6698ef2-c594-4859-ba6d-7d76127c819d/operator/0.log" Dec 01 18:21:43 crc kubenswrapper[4868]: I1201 18:21:43.386534 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-97g6w_d87f1b79-a4fb-4cfa-86d7-823ceedcf913/kube-rbac-proxy/0.log" Dec 01 18:21:43 crc kubenswrapper[4868]: I1201 18:21:43.428548 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-97g6w_d87f1b79-a4fb-4cfa-86d7-823ceedcf913/manager/0.log" Dec 01 18:21:43 crc kubenswrapper[4868]: I1201 18:21:43.563581 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-mmjrx_d5d35af7-2b67-42cd-ae7c-d73f105da292/kube-rbac-proxy/0.log" Dec 01 18:21:43 crc kubenswrapper[4868]: I1201 18:21:43.676884 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-mmjrx_d5d35af7-2b67-42cd-ae7c-d73f105da292/manager/0.log" Dec 01 18:21:43 crc kubenswrapper[4868]: I1201 18:21:43.682338 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-blw9f_32e03cbc-4f29-49ee-8558-3f8950b0a383/operator/0.log" Dec 01 18:21:43 crc kubenswrapper[4868]: I1201 18:21:43.900188 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-g56bv_58e2eaed-6851-4568-a604-397a8944da7f/kube-rbac-proxy/0.log" Dec 01 18:21:43 crc kubenswrapper[4868]: I1201 18:21:43.977711 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-g56bv_58e2eaed-6851-4568-a604-397a8944da7f/manager/0.log" Dec 01 18:21:44 crc kubenswrapper[4868]: I1201 18:21:44.039387 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-755b486884-bvzjt_515c8a14-67ad-4c4c-8b3e-9433997fbdb3/manager/0.log" Dec 01 18:21:44 crc kubenswrapper[4868]: I1201 18:21:44.187127 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-l4hqj_8df8451f-54fc-4553-834b-37839d4c1807/kube-rbac-proxy/0.log" Dec 01 18:21:44 crc kubenswrapper[4868]: I1201 18:21:44.196229 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-l4hqj_8df8451f-54fc-4553-834b-37839d4c1807/manager/0.log" Dec 01 18:21:44 crc kubenswrapper[4868]: I1201 18:21:44.200388 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-d2jx2_f1f47ec9-ea94-4fd5-b113-baa29c640369/kube-rbac-proxy/0.log" Dec 01 18:21:44 crc kubenswrapper[4868]: I1201 18:21:44.291308 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-d2jx2_f1f47ec9-ea94-4fd5-b113-baa29c640369/manager/0.log" Dec 01 18:21:44 crc kubenswrapper[4868]: I1201 18:21:44.362647 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-b78w5_fc4efb42-45ab-4341-b6f2-d95497223de3/kube-rbac-proxy/0.log" Dec 01 18:21:44 crc kubenswrapper[4868]: I1201 18:21:44.402087 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-b78w5_fc4efb42-45ab-4341-b6f2-d95497223de3/manager/0.log" Dec 01 18:21:57 crc kubenswrapper[4868]: I1201 18:21:57.172477 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:21:57 crc kubenswrapper[4868]: E1201 18:21:57.174356 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:22:03 crc kubenswrapper[4868]: I1201 18:22:03.385402 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-vw9z9_2bf15be5-1380-4b58-934d-7793e22541b4/control-plane-machine-set-operator/0.log" Dec 01 18:22:03 crc kubenswrapper[4868]: I1201 18:22:03.576846 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qw86w_be098182-dd51-4a19-920a-a797a175124a/kube-rbac-proxy/0.log" Dec 01 18:22:03 crc kubenswrapper[4868]: I1201 18:22:03.616137 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qw86w_be098182-dd51-4a19-920a-a797a175124a/machine-api-operator/0.log" Dec 01 18:22:10 crc kubenswrapper[4868]: I1201 18:22:10.172155 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:22:10 crc kubenswrapper[4868]: E1201 18:22:10.172875 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:22:16 crc kubenswrapper[4868]: I1201 18:22:16.230694 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-f6tcg_958205b1-cd35-41d8-9e21-b30cc80a2489/cert-manager-controller/0.log" Dec 01 18:22:16 crc kubenswrapper[4868]: I1201 18:22:16.359040 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-j2dnr_8057b33e-2417-43f9-aab1-7f7db221d051/cert-manager-cainjector/0.log" Dec 01 18:22:16 crc kubenswrapper[4868]: I1201 18:22:16.414156 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-qgmqm_3858baab-f6b9-42a5-b7ad-2c93a8e98fb6/cert-manager-webhook/0.log" Dec 01 18:22:24 crc kubenswrapper[4868]: I1201 18:22:24.174065 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:22:24 crc kubenswrapper[4868]: E1201 18:22:24.174750 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:22:29 crc kubenswrapper[4868]: I1201 18:22:29.661120 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-xsj4b_fa121055-74c0-44aa-8401-3a8476c7db4c/nmstate-console-plugin/0.log" Dec 01 18:22:29 crc kubenswrapper[4868]: I1201 18:22:29.879879 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-67rpb_d04c81f8-7543-44b9-abc9-b0c95346cd56/nmstate-handler/0.log" Dec 01 18:22:30 crc kubenswrapper[4868]: I1201 18:22:30.005672 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-j7lcc_ed23de21-61e1-463e-9a8e-5a8ba51cc727/kube-rbac-proxy/0.log" Dec 01 18:22:30 crc kubenswrapper[4868]: I1201 18:22:30.053789 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-j7lcc_ed23de21-61e1-463e-9a8e-5a8ba51cc727/nmstate-metrics/0.log" Dec 01 18:22:30 crc kubenswrapper[4868]: I1201 18:22:30.209980 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-rb8wl_13c419e8-4943-4ee5-8d1c-5dd5051c81db/nmstate-operator/0.log" Dec 01 18:22:30 crc kubenswrapper[4868]: I1201 18:22:30.293135 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-kjwkw_f1784373-a1b0-4ca1-8c08-cf01fbc83b72/nmstate-webhook/0.log" Dec 01 18:22:36 crc kubenswrapper[4868]: I1201 18:22:36.173032 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:22:36 crc kubenswrapper[4868]: I1201 18:22:36.901909 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"d96440aad2ff00174b44608542db90d83651ffced39d42cc58197244d3a95475"} Dec 01 18:22:45 crc kubenswrapper[4868]: I1201 18:22:45.182971 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-nvnm5_5536b1ca-ecee-4f69-9884-539b000d00f0/kube-rbac-proxy/0.log" Dec 01 18:22:45 crc kubenswrapper[4868]: I1201 18:22:45.267479 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-nvnm5_5536b1ca-ecee-4f69-9884-539b000d00f0/controller/0.log" Dec 01 18:22:45 crc kubenswrapper[4868]: I1201 18:22:45.434463 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-frr-files/0.log" Dec 01 18:22:45 crc kubenswrapper[4868]: I1201 18:22:45.606770 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-frr-files/0.log" Dec 01 18:22:45 crc kubenswrapper[4868]: I1201 18:22:45.607053 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-reloader/0.log" Dec 01 18:22:45 crc kubenswrapper[4868]: I1201 18:22:45.654831 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-reloader/0.log" Dec 01 18:22:45 crc kubenswrapper[4868]: I1201 18:22:45.668955 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-metrics/0.log" Dec 01 18:22:45 crc kubenswrapper[4868]: I1201 18:22:45.839195 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-reloader/0.log" Dec 01 18:22:45 crc kubenswrapper[4868]: I1201 18:22:45.844693 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-frr-files/0.log" Dec 01 18:22:45 crc kubenswrapper[4868]: I1201 18:22:45.882212 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-metrics/0.log" Dec 01 18:22:45 crc kubenswrapper[4868]: I1201 18:22:45.883241 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-metrics/0.log" Dec 01 18:22:46 crc kubenswrapper[4868]: I1201 18:22:46.119101 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-metrics/0.log" Dec 01 18:22:46 crc kubenswrapper[4868]: I1201 18:22:46.281841 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-reloader/0.log" Dec 01 18:22:46 crc kubenswrapper[4868]: I1201 18:22:46.306561 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-frr-files/0.log" Dec 01 18:22:46 crc kubenswrapper[4868]: I1201 18:22:46.321050 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/controller/0.log" Dec 01 18:22:46 crc kubenswrapper[4868]: I1201 18:22:46.507904 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/frr-metrics/0.log" Dec 01 18:22:46 crc kubenswrapper[4868]: I1201 18:22:46.548103 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/kube-rbac-proxy/0.log" Dec 01 18:22:46 crc kubenswrapper[4868]: I1201 18:22:46.674748 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/kube-rbac-proxy-frr/0.log" Dec 01 18:22:46 crc kubenswrapper[4868]: I1201 18:22:46.786907 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/reloader/0.log" Dec 01 18:22:46 crc kubenswrapper[4868]: I1201 18:22:46.852025 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-54t6t_e22c99ca-b516-4e36-bc9b-ba5355170fd8/frr-k8s-webhook-server/0.log" Dec 01 18:22:47 crc kubenswrapper[4868]: I1201 18:22:47.099647 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-779b644dbb-f6925_0724b3ed-ae00-455d-9f48-94cd89bf56d9/manager/0.log" Dec 01 18:22:47 crc kubenswrapper[4868]: I1201 18:22:47.258344 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-85f6846bc9-2b4v6_f0ad7ab5-3180-42c7-90a1-7a75fac7a61d/webhook-server/0.log" Dec 01 18:22:47 crc kubenswrapper[4868]: I1201 18:22:47.294227 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xfljl_22663c8b-3ec6-453c-bb25-ca8d92576d07/kube-rbac-proxy/0.log" Dec 01 18:22:47 crc kubenswrapper[4868]: I1201 18:22:47.734786 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/frr/0.log" Dec 01 18:22:47 crc kubenswrapper[4868]: I1201 18:22:47.849269 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xfljl_22663c8b-3ec6-453c-bb25-ca8d92576d07/speaker/0.log" Dec 01 18:23:00 crc kubenswrapper[4868]: I1201 18:23:00.229228 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6_e1d4847d-1891-4e4c-9d68-04181da92d62/util/0.log" Dec 01 18:23:00 crc kubenswrapper[4868]: I1201 18:23:00.367837 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6_e1d4847d-1891-4e4c-9d68-04181da92d62/pull/0.log" Dec 01 18:23:00 crc kubenswrapper[4868]: I1201 18:23:00.371630 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6_e1d4847d-1891-4e4c-9d68-04181da92d62/pull/0.log" Dec 01 18:23:00 crc kubenswrapper[4868]: I1201 18:23:00.405204 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6_e1d4847d-1891-4e4c-9d68-04181da92d62/util/0.log" Dec 01 18:23:00 crc kubenswrapper[4868]: I1201 18:23:00.581105 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6_e1d4847d-1891-4e4c-9d68-04181da92d62/pull/0.log" Dec 01 18:23:00 crc kubenswrapper[4868]: I1201 18:23:00.605214 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6_e1d4847d-1891-4e4c-9d68-04181da92d62/util/0.log" Dec 01 18:23:00 crc kubenswrapper[4868]: I1201 18:23:00.618648 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6_e1d4847d-1891-4e4c-9d68-04181da92d62/extract/0.log" Dec 01 18:23:00 crc kubenswrapper[4868]: I1201 18:23:00.779689 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv_22daded3-8016-41cc-bc8b-0dfc2ac3c093/util/0.log" Dec 01 18:23:01 crc kubenswrapper[4868]: I1201 18:23:01.028167 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv_22daded3-8016-41cc-bc8b-0dfc2ac3c093/util/0.log" Dec 01 18:23:01 crc kubenswrapper[4868]: I1201 18:23:01.064480 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv_22daded3-8016-41cc-bc8b-0dfc2ac3c093/pull/0.log" Dec 01 18:23:01 crc kubenswrapper[4868]: I1201 18:23:01.073352 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv_22daded3-8016-41cc-bc8b-0dfc2ac3c093/pull/0.log" Dec 01 18:23:01 crc kubenswrapper[4868]: I1201 18:23:01.266977 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv_22daded3-8016-41cc-bc8b-0dfc2ac3c093/pull/0.log" Dec 01 18:23:01 crc kubenswrapper[4868]: I1201 18:23:01.271360 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv_22daded3-8016-41cc-bc8b-0dfc2ac3c093/extract/0.log" Dec 01 18:23:01 crc kubenswrapper[4868]: I1201 18:23:01.274838 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv_22daded3-8016-41cc-bc8b-0dfc2ac3c093/util/0.log" Dec 01 18:23:01 crc kubenswrapper[4868]: I1201 18:23:01.539090 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-znxj6_2faafa31-ca8e-4da4-a1b4-e3d2709b4648/extract-utilities/0.log" Dec 01 18:23:01 crc kubenswrapper[4868]: I1201 18:23:01.690879 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-znxj6_2faafa31-ca8e-4da4-a1b4-e3d2709b4648/extract-utilities/0.log" Dec 01 18:23:01 crc kubenswrapper[4868]: I1201 18:23:01.735229 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-znxj6_2faafa31-ca8e-4da4-a1b4-e3d2709b4648/extract-content/0.log" Dec 01 18:23:01 crc kubenswrapper[4868]: I1201 18:23:01.748478 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-znxj6_2faafa31-ca8e-4da4-a1b4-e3d2709b4648/extract-content/0.log" Dec 01 18:23:01 crc kubenswrapper[4868]: I1201 18:23:01.897226 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-znxj6_2faafa31-ca8e-4da4-a1b4-e3d2709b4648/extract-utilities/0.log" Dec 01 18:23:01 crc kubenswrapper[4868]: I1201 18:23:01.915742 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-znxj6_2faafa31-ca8e-4da4-a1b4-e3d2709b4648/extract-content/0.log" Dec 01 18:23:02 crc kubenswrapper[4868]: I1201 18:23:02.093830 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fpxc7_ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a/extract-utilities/0.log" Dec 01 18:23:02 crc kubenswrapper[4868]: I1201 18:23:02.403276 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fpxc7_ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a/extract-content/0.log" Dec 01 18:23:02 crc kubenswrapper[4868]: I1201 18:23:02.424229 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fpxc7_ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a/extract-utilities/0.log" Dec 01 18:23:02 crc kubenswrapper[4868]: I1201 18:23:02.451170 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fpxc7_ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a/extract-content/0.log" Dec 01 18:23:02 crc kubenswrapper[4868]: I1201 18:23:02.615193 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-znxj6_2faafa31-ca8e-4da4-a1b4-e3d2709b4648/registry-server/0.log" Dec 01 18:23:02 crc kubenswrapper[4868]: I1201 18:23:02.663462 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fpxc7_ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a/extract-content/0.log" Dec 01 18:23:02 crc kubenswrapper[4868]: I1201 18:23:02.714997 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fpxc7_ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a/extract-utilities/0.log" Dec 01 18:23:02 crc kubenswrapper[4868]: I1201 18:23:02.871744 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4nfp5_0c043a8c-1af1-4aa0-99f9-0f6ebe27238a/marketplace-operator/0.log" Dec 01 18:23:02 crc kubenswrapper[4868]: I1201 18:23:02.917604 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fpxc7_ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a/registry-server/0.log" Dec 01 18:23:03 crc kubenswrapper[4868]: I1201 18:23:03.062104 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9kf2c_5bff6f35-58ce-48e1-ae66-e29973c57a3f/extract-utilities/0.log" Dec 01 18:23:03 crc kubenswrapper[4868]: I1201 18:23:03.227225 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9kf2c_5bff6f35-58ce-48e1-ae66-e29973c57a3f/extract-utilities/0.log" Dec 01 18:23:03 crc kubenswrapper[4868]: I1201 18:23:03.241668 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9kf2c_5bff6f35-58ce-48e1-ae66-e29973c57a3f/extract-content/0.log" Dec 01 18:23:03 crc kubenswrapper[4868]: I1201 18:23:03.249387 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9kf2c_5bff6f35-58ce-48e1-ae66-e29973c57a3f/extract-content/0.log" Dec 01 18:23:03 crc kubenswrapper[4868]: I1201 18:23:03.422827 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9kf2c_5bff6f35-58ce-48e1-ae66-e29973c57a3f/extract-utilities/0.log" Dec 01 18:23:03 crc kubenswrapper[4868]: I1201 18:23:03.509704 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9kf2c_5bff6f35-58ce-48e1-ae66-e29973c57a3f/extract-content/0.log" Dec 01 18:23:03 crc kubenswrapper[4868]: I1201 18:23:03.538478 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9kf2c_5bff6f35-58ce-48e1-ae66-e29973c57a3f/registry-server/0.log" Dec 01 18:23:03 crc kubenswrapper[4868]: I1201 18:23:03.625041 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rcxdx_a904bedc-9d01-4a00-82c1-6f2826a33d81/extract-utilities/0.log" Dec 01 18:23:03 crc kubenswrapper[4868]: I1201 18:23:03.812347 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rcxdx_a904bedc-9d01-4a00-82c1-6f2826a33d81/extract-utilities/0.log" Dec 01 18:23:03 crc kubenswrapper[4868]: I1201 18:23:03.829848 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rcxdx_a904bedc-9d01-4a00-82c1-6f2826a33d81/extract-content/0.log" Dec 01 18:23:03 crc kubenswrapper[4868]: I1201 18:23:03.831814 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rcxdx_a904bedc-9d01-4a00-82c1-6f2826a33d81/extract-content/0.log" Dec 01 18:23:04 crc kubenswrapper[4868]: I1201 18:23:04.021583 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rcxdx_a904bedc-9d01-4a00-82c1-6f2826a33d81/extract-content/0.log" Dec 01 18:23:04 crc kubenswrapper[4868]: I1201 18:23:04.022872 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rcxdx_a904bedc-9d01-4a00-82c1-6f2826a33d81/extract-utilities/0.log" Dec 01 18:23:04 crc kubenswrapper[4868]: I1201 18:23:04.411730 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rcxdx_a904bedc-9d01-4a00-82c1-6f2826a33d81/registry-server/0.log" Dec 01 18:23:26 crc kubenswrapper[4868]: E1201 18:23:26.348852 4868 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.103:34406->38.102.83.103:33157: write tcp 38.102.83.103:34406->38.102.83.103:33157: write: broken pipe Dec 01 18:24:42 crc kubenswrapper[4868]: I1201 18:24:42.101855 4868 generic.go:334] "Generic (PLEG): container finished" podID="2a60cf14-31c5-461f-816f-4f97fd2999a5" containerID="417958dfa867323de1112a7435dd8a2354512c6bff43f60bcfdb99f869f9864b" exitCode=0 Dec 01 18:24:42 crc kubenswrapper[4868]: I1201 18:24:42.101920 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5kwbg/must-gather-plcx8" event={"ID":"2a60cf14-31c5-461f-816f-4f97fd2999a5","Type":"ContainerDied","Data":"417958dfa867323de1112a7435dd8a2354512c6bff43f60bcfdb99f869f9864b"} Dec 01 18:24:42 crc kubenswrapper[4868]: I1201 18:24:42.102867 4868 scope.go:117] "RemoveContainer" containerID="417958dfa867323de1112a7435dd8a2354512c6bff43f60bcfdb99f869f9864b" Dec 01 18:24:42 crc kubenswrapper[4868]: I1201 18:24:42.826521 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5kwbg_must-gather-plcx8_2a60cf14-31c5-461f-816f-4f97fd2999a5/gather/0.log" Dec 01 18:24:50 crc kubenswrapper[4868]: I1201 18:24:50.369700 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5kwbg/must-gather-plcx8"] Dec 01 18:24:50 crc kubenswrapper[4868]: I1201 18:24:50.370542 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-5kwbg/must-gather-plcx8" podUID="2a60cf14-31c5-461f-816f-4f97fd2999a5" containerName="copy" containerID="cri-o://07b747309324d285cf57258ad3a4c942bd9c38821e7bd8cb4a434eecf643f992" gracePeriod=2 Dec 01 18:24:50 crc kubenswrapper[4868]: I1201 18:24:50.378000 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5kwbg/must-gather-plcx8"] Dec 01 18:24:50 crc kubenswrapper[4868]: I1201 18:24:50.753729 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5kwbg_must-gather-plcx8_2a60cf14-31c5-461f-816f-4f97fd2999a5/copy/0.log" Dec 01 18:24:50 crc kubenswrapper[4868]: I1201 18:24:50.754350 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/must-gather-plcx8" Dec 01 18:24:50 crc kubenswrapper[4868]: I1201 18:24:50.793739 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2a60cf14-31c5-461f-816f-4f97fd2999a5-must-gather-output\") pod \"2a60cf14-31c5-461f-816f-4f97fd2999a5\" (UID: \"2a60cf14-31c5-461f-816f-4f97fd2999a5\") " Dec 01 18:24:50 crc kubenswrapper[4868]: I1201 18:24:50.793803 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7z7g\" (UniqueName: \"kubernetes.io/projected/2a60cf14-31c5-461f-816f-4f97fd2999a5-kube-api-access-c7z7g\") pod \"2a60cf14-31c5-461f-816f-4f97fd2999a5\" (UID: \"2a60cf14-31c5-461f-816f-4f97fd2999a5\") " Dec 01 18:24:50 crc kubenswrapper[4868]: I1201 18:24:50.806242 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a60cf14-31c5-461f-816f-4f97fd2999a5-kube-api-access-c7z7g" (OuterVolumeSpecName: "kube-api-access-c7z7g") pod "2a60cf14-31c5-461f-816f-4f97fd2999a5" (UID: "2a60cf14-31c5-461f-816f-4f97fd2999a5"). InnerVolumeSpecName "kube-api-access-c7z7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:24:50 crc kubenswrapper[4868]: I1201 18:24:50.896085 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7z7g\" (UniqueName: \"kubernetes.io/projected/2a60cf14-31c5-461f-816f-4f97fd2999a5-kube-api-access-c7z7g\") on node \"crc\" DevicePath \"\"" Dec 01 18:24:50 crc kubenswrapper[4868]: I1201 18:24:50.962481 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a60cf14-31c5-461f-816f-4f97fd2999a5-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2a60cf14-31c5-461f-816f-4f97fd2999a5" (UID: "2a60cf14-31c5-461f-816f-4f97fd2999a5"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:24:50 crc kubenswrapper[4868]: I1201 18:24:50.998243 4868 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2a60cf14-31c5-461f-816f-4f97fd2999a5-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 01 18:24:51 crc kubenswrapper[4868]: I1201 18:24:51.178494 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5kwbg_must-gather-plcx8_2a60cf14-31c5-461f-816f-4f97fd2999a5/copy/0.log" Dec 01 18:24:51 crc kubenswrapper[4868]: I1201 18:24:51.178886 4868 generic.go:334] "Generic (PLEG): container finished" podID="2a60cf14-31c5-461f-816f-4f97fd2999a5" containerID="07b747309324d285cf57258ad3a4c942bd9c38821e7bd8cb4a434eecf643f992" exitCode=143 Dec 01 18:24:51 crc kubenswrapper[4868]: I1201 18:24:51.178936 4868 scope.go:117] "RemoveContainer" containerID="07b747309324d285cf57258ad3a4c942bd9c38821e7bd8cb4a434eecf643f992" Dec 01 18:24:51 crc kubenswrapper[4868]: I1201 18:24:51.179004 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5kwbg/must-gather-plcx8" Dec 01 18:24:51 crc kubenswrapper[4868]: I1201 18:24:51.207860 4868 scope.go:117] "RemoveContainer" containerID="417958dfa867323de1112a7435dd8a2354512c6bff43f60bcfdb99f869f9864b" Dec 01 18:24:51 crc kubenswrapper[4868]: I1201 18:24:51.256041 4868 scope.go:117] "RemoveContainer" containerID="07b747309324d285cf57258ad3a4c942bd9c38821e7bd8cb4a434eecf643f992" Dec 01 18:24:51 crc kubenswrapper[4868]: E1201 18:24:51.256517 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07b747309324d285cf57258ad3a4c942bd9c38821e7bd8cb4a434eecf643f992\": container with ID starting with 07b747309324d285cf57258ad3a4c942bd9c38821e7bd8cb4a434eecf643f992 not found: ID does not exist" containerID="07b747309324d285cf57258ad3a4c942bd9c38821e7bd8cb4a434eecf643f992" Dec 01 18:24:51 crc kubenswrapper[4868]: I1201 18:24:51.256559 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07b747309324d285cf57258ad3a4c942bd9c38821e7bd8cb4a434eecf643f992"} err="failed to get container status \"07b747309324d285cf57258ad3a4c942bd9c38821e7bd8cb4a434eecf643f992\": rpc error: code = NotFound desc = could not find container \"07b747309324d285cf57258ad3a4c942bd9c38821e7bd8cb4a434eecf643f992\": container with ID starting with 07b747309324d285cf57258ad3a4c942bd9c38821e7bd8cb4a434eecf643f992 not found: ID does not exist" Dec 01 18:24:51 crc kubenswrapper[4868]: I1201 18:24:51.256589 4868 scope.go:117] "RemoveContainer" containerID="417958dfa867323de1112a7435dd8a2354512c6bff43f60bcfdb99f869f9864b" Dec 01 18:24:51 crc kubenswrapper[4868]: E1201 18:24:51.256899 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"417958dfa867323de1112a7435dd8a2354512c6bff43f60bcfdb99f869f9864b\": container with ID starting with 417958dfa867323de1112a7435dd8a2354512c6bff43f60bcfdb99f869f9864b not found: ID does not exist" containerID="417958dfa867323de1112a7435dd8a2354512c6bff43f60bcfdb99f869f9864b" Dec 01 18:24:51 crc kubenswrapper[4868]: I1201 18:24:51.256932 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"417958dfa867323de1112a7435dd8a2354512c6bff43f60bcfdb99f869f9864b"} err="failed to get container status \"417958dfa867323de1112a7435dd8a2354512c6bff43f60bcfdb99f869f9864b\": rpc error: code = NotFound desc = could not find container \"417958dfa867323de1112a7435dd8a2354512c6bff43f60bcfdb99f869f9864b\": container with ID starting with 417958dfa867323de1112a7435dd8a2354512c6bff43f60bcfdb99f869f9864b not found: ID does not exist" Dec 01 18:24:52 crc kubenswrapper[4868]: I1201 18:24:52.185029 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a60cf14-31c5-461f-816f-4f97fd2999a5" path="/var/lib/kubelet/pods/2a60cf14-31c5-461f-816f-4f97fd2999a5/volumes" Dec 01 18:24:55 crc kubenswrapper[4868]: I1201 18:24:55.905309 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:24:55 crc kubenswrapper[4868]: I1201 18:24:55.906221 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:25:25 crc kubenswrapper[4868]: I1201 18:25:25.905178 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:25:25 crc kubenswrapper[4868]: I1201 18:25:25.906784 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:25:55 crc kubenswrapper[4868]: I1201 18:25:55.904985 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:25:55 crc kubenswrapper[4868]: I1201 18:25:55.905510 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:25:55 crc kubenswrapper[4868]: I1201 18:25:55.905560 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 18:25:55 crc kubenswrapper[4868]: I1201 18:25:55.906118 4868 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d96440aad2ff00174b44608542db90d83651ffced39d42cc58197244d3a95475"} pod="openshift-machine-config-operator/machine-config-daemon-njgpn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 18:25:55 crc kubenswrapper[4868]: I1201 18:25:55.906177 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" containerID="cri-o://d96440aad2ff00174b44608542db90d83651ffced39d42cc58197244d3a95475" gracePeriod=600 Dec 01 18:25:56 crc kubenswrapper[4868]: I1201 18:25:56.749305 4868 generic.go:334] "Generic (PLEG): container finished" podID="ec4e08bf-df56-4490-8e79-a4175c727405" containerID="d96440aad2ff00174b44608542db90d83651ffced39d42cc58197244d3a95475" exitCode=0 Dec 01 18:25:56 crc kubenswrapper[4868]: I1201 18:25:56.749382 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerDied","Data":"d96440aad2ff00174b44608542db90d83651ffced39d42cc58197244d3a95475"} Dec 01 18:25:56 crc kubenswrapper[4868]: I1201 18:25:56.749548 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f"} Dec 01 18:25:56 crc kubenswrapper[4868]: I1201 18:25:56.749566 4868 scope.go:117] "RemoveContainer" containerID="9732fa4cf0e72267502e2ace88222f0192a64413d5293caf0de3dad6f36a8bc6" Dec 01 18:26:42 crc kubenswrapper[4868]: I1201 18:26:42.063182 4868 scope.go:117] "RemoveContainer" containerID="bc4ed594d91529456ebc5cb17b988e45ef10d247116675aef7bd059e1c3b4378" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.231442 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gd4nc"] Dec 01 18:27:38 crc kubenswrapper[4868]: E1201 18:27:38.232795 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbd7e01-7755-4477-a7ce-d976040ecd2f" containerName="extract-utilities" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.232819 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbd7e01-7755-4477-a7ce-d976040ecd2f" containerName="extract-utilities" Dec 01 18:27:38 crc kubenswrapper[4868]: E1201 18:27:38.232866 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a60cf14-31c5-461f-816f-4f97fd2999a5" containerName="copy" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.232878 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a60cf14-31c5-461f-816f-4f97fd2999a5" containerName="copy" Dec 01 18:27:38 crc kubenswrapper[4868]: E1201 18:27:38.232897 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a60cf14-31c5-461f-816f-4f97fd2999a5" containerName="gather" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.232912 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a60cf14-31c5-461f-816f-4f97fd2999a5" containerName="gather" Dec 01 18:27:38 crc kubenswrapper[4868]: E1201 18:27:38.232935 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbd7e01-7755-4477-a7ce-d976040ecd2f" containerName="extract-content" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.232973 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbd7e01-7755-4477-a7ce-d976040ecd2f" containerName="extract-content" Dec 01 18:27:38 crc kubenswrapper[4868]: E1201 18:27:38.233006 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbd7e01-7755-4477-a7ce-d976040ecd2f" containerName="registry-server" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.233017 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbd7e01-7755-4477-a7ce-d976040ecd2f" containerName="registry-server" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.233311 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dbd7e01-7755-4477-a7ce-d976040ecd2f" containerName="registry-server" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.233350 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a60cf14-31c5-461f-816f-4f97fd2999a5" containerName="gather" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.233365 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a60cf14-31c5-461f-816f-4f97fd2999a5" containerName="copy" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.235648 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.242616 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gd4nc"] Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.373558 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bf44ca1-5a2e-4150-b255-f104d6752454-catalog-content\") pod \"redhat-marketplace-gd4nc\" (UID: \"0bf44ca1-5a2e-4150-b255-f104d6752454\") " pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.373640 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bf44ca1-5a2e-4150-b255-f104d6752454-utilities\") pod \"redhat-marketplace-gd4nc\" (UID: \"0bf44ca1-5a2e-4150-b255-f104d6752454\") " pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.373742 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctfd4\" (UniqueName: \"kubernetes.io/projected/0bf44ca1-5a2e-4150-b255-f104d6752454-kube-api-access-ctfd4\") pod \"redhat-marketplace-gd4nc\" (UID: \"0bf44ca1-5a2e-4150-b255-f104d6752454\") " pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.475813 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bf44ca1-5a2e-4150-b255-f104d6752454-catalog-content\") pod \"redhat-marketplace-gd4nc\" (UID: \"0bf44ca1-5a2e-4150-b255-f104d6752454\") " pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.476294 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bf44ca1-5a2e-4150-b255-f104d6752454-utilities\") pod \"redhat-marketplace-gd4nc\" (UID: \"0bf44ca1-5a2e-4150-b255-f104d6752454\") " pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.476437 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctfd4\" (UniqueName: \"kubernetes.io/projected/0bf44ca1-5a2e-4150-b255-f104d6752454-kube-api-access-ctfd4\") pod \"redhat-marketplace-gd4nc\" (UID: \"0bf44ca1-5a2e-4150-b255-f104d6752454\") " pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.476550 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bf44ca1-5a2e-4150-b255-f104d6752454-catalog-content\") pod \"redhat-marketplace-gd4nc\" (UID: \"0bf44ca1-5a2e-4150-b255-f104d6752454\") " pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.476879 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bf44ca1-5a2e-4150-b255-f104d6752454-utilities\") pod \"redhat-marketplace-gd4nc\" (UID: \"0bf44ca1-5a2e-4150-b255-f104d6752454\") " pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.500545 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctfd4\" (UniqueName: \"kubernetes.io/projected/0bf44ca1-5a2e-4150-b255-f104d6752454-kube-api-access-ctfd4\") pod \"redhat-marketplace-gd4nc\" (UID: \"0bf44ca1-5a2e-4150-b255-f104d6752454\") " pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:38 crc kubenswrapper[4868]: I1201 18:27:38.566214 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:39 crc kubenswrapper[4868]: I1201 18:27:39.127337 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gd4nc"] Dec 01 18:27:39 crc kubenswrapper[4868]: I1201 18:27:39.764003 4868 generic.go:334] "Generic (PLEG): container finished" podID="0bf44ca1-5a2e-4150-b255-f104d6752454" containerID="a4c05f8b0ccdad667ef20ae3299eceaf676bc7b20d80c81aa429e611edbd3e51" exitCode=0 Dec 01 18:27:39 crc kubenswrapper[4868]: I1201 18:27:39.764066 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gd4nc" event={"ID":"0bf44ca1-5a2e-4150-b255-f104d6752454","Type":"ContainerDied","Data":"a4c05f8b0ccdad667ef20ae3299eceaf676bc7b20d80c81aa429e611edbd3e51"} Dec 01 18:27:39 crc kubenswrapper[4868]: I1201 18:27:39.764108 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gd4nc" event={"ID":"0bf44ca1-5a2e-4150-b255-f104d6752454","Type":"ContainerStarted","Data":"784dff73d038f3565d287a64b2adc7f61161da4e65cf0230f968fbc3113806dc"} Dec 01 18:27:39 crc kubenswrapper[4868]: I1201 18:27:39.766270 4868 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 18:27:40 crc kubenswrapper[4868]: I1201 18:27:40.775747 4868 generic.go:334] "Generic (PLEG): container finished" podID="0bf44ca1-5a2e-4150-b255-f104d6752454" containerID="349e991ffc7ef1c369dbe6da38e9ea53a6c62948191adbb1f731af7d5d5bd884" exitCode=0 Dec 01 18:27:40 crc kubenswrapper[4868]: I1201 18:27:40.775983 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gd4nc" event={"ID":"0bf44ca1-5a2e-4150-b255-f104d6752454","Type":"ContainerDied","Data":"349e991ffc7ef1c369dbe6da38e9ea53a6c62948191adbb1f731af7d5d5bd884"} Dec 01 18:27:41 crc kubenswrapper[4868]: I1201 18:27:41.787160 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gd4nc" event={"ID":"0bf44ca1-5a2e-4150-b255-f104d6752454","Type":"ContainerStarted","Data":"572fec971502506024c948a95e29b602877b68cb2dedd4885c8ab8cd347220ad"} Dec 01 18:27:41 crc kubenswrapper[4868]: I1201 18:27:41.811935 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gd4nc" podStartSLOduration=2.027891998 podStartE2EDuration="3.811913149s" podCreationTimestamp="2025-12-01 18:27:38 +0000 UTC" firstStartedPulling="2025-12-01 18:27:39.766048711 +0000 UTC m=+3732.137159122" lastFinishedPulling="2025-12-01 18:27:41.550069862 +0000 UTC m=+3733.921180273" observedRunningTime="2025-12-01 18:27:41.806213145 +0000 UTC m=+3734.177323566" watchObservedRunningTime="2025-12-01 18:27:41.811913149 +0000 UTC m=+3734.183023560" Dec 01 18:27:42 crc kubenswrapper[4868]: I1201 18:27:42.703852 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vddj9/must-gather-k8hct"] Dec 01 18:27:42 crc kubenswrapper[4868]: I1201 18:27:42.743597 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-vddj9/must-gather-k8hct"] Dec 01 18:27:42 crc kubenswrapper[4868]: I1201 18:27:42.743730 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/must-gather-k8hct" Dec 01 18:27:42 crc kubenswrapper[4868]: I1201 18:27:42.745924 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-vddj9"/"openshift-service-ca.crt" Dec 01 18:27:42 crc kubenswrapper[4868]: I1201 18:27:42.746168 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-vddj9"/"default-dockercfg-jlzlq" Dec 01 18:27:42 crc kubenswrapper[4868]: I1201 18:27:42.748175 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-vddj9"/"kube-root-ca.crt" Dec 01 18:27:42 crc kubenswrapper[4868]: I1201 18:27:42.863250 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1ef4472c-7831-4220-a58e-efb68fa8d78e-must-gather-output\") pod \"must-gather-k8hct\" (UID: \"1ef4472c-7831-4220-a58e-efb68fa8d78e\") " pod="openshift-must-gather-vddj9/must-gather-k8hct" Dec 01 18:27:42 crc kubenswrapper[4868]: I1201 18:27:42.863540 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmckq\" (UniqueName: \"kubernetes.io/projected/1ef4472c-7831-4220-a58e-efb68fa8d78e-kube-api-access-fmckq\") pod \"must-gather-k8hct\" (UID: \"1ef4472c-7831-4220-a58e-efb68fa8d78e\") " pod="openshift-must-gather-vddj9/must-gather-k8hct" Dec 01 18:27:42 crc kubenswrapper[4868]: I1201 18:27:42.965469 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1ef4472c-7831-4220-a58e-efb68fa8d78e-must-gather-output\") pod \"must-gather-k8hct\" (UID: \"1ef4472c-7831-4220-a58e-efb68fa8d78e\") " pod="openshift-must-gather-vddj9/must-gather-k8hct" Dec 01 18:27:42 crc kubenswrapper[4868]: I1201 18:27:42.965549 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmckq\" (UniqueName: \"kubernetes.io/projected/1ef4472c-7831-4220-a58e-efb68fa8d78e-kube-api-access-fmckq\") pod \"must-gather-k8hct\" (UID: \"1ef4472c-7831-4220-a58e-efb68fa8d78e\") " pod="openshift-must-gather-vddj9/must-gather-k8hct" Dec 01 18:27:42 crc kubenswrapper[4868]: I1201 18:27:42.966053 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1ef4472c-7831-4220-a58e-efb68fa8d78e-must-gather-output\") pod \"must-gather-k8hct\" (UID: \"1ef4472c-7831-4220-a58e-efb68fa8d78e\") " pod="openshift-must-gather-vddj9/must-gather-k8hct" Dec 01 18:27:42 crc kubenswrapper[4868]: I1201 18:27:42.990586 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmckq\" (UniqueName: \"kubernetes.io/projected/1ef4472c-7831-4220-a58e-efb68fa8d78e-kube-api-access-fmckq\") pod \"must-gather-k8hct\" (UID: \"1ef4472c-7831-4220-a58e-efb68fa8d78e\") " pod="openshift-must-gather-vddj9/must-gather-k8hct" Dec 01 18:27:43 crc kubenswrapper[4868]: I1201 18:27:43.077855 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/must-gather-k8hct" Dec 01 18:27:43 crc kubenswrapper[4868]: I1201 18:27:43.509594 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-vddj9/must-gather-k8hct"] Dec 01 18:27:43 crc kubenswrapper[4868]: W1201 18:27:43.515373 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ef4472c_7831_4220_a58e_efb68fa8d78e.slice/crio-47cdedfebf40b7e4bdee47ce9157f6ee3b7141bf7b4b5d7d372645084f412a1a WatchSource:0}: Error finding container 47cdedfebf40b7e4bdee47ce9157f6ee3b7141bf7b4b5d7d372645084f412a1a: Status 404 returned error can't find the container with id 47cdedfebf40b7e4bdee47ce9157f6ee3b7141bf7b4b5d7d372645084f412a1a Dec 01 18:27:43 crc kubenswrapper[4868]: I1201 18:27:43.807130 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vddj9/must-gather-k8hct" event={"ID":"1ef4472c-7831-4220-a58e-efb68fa8d78e","Type":"ContainerStarted","Data":"23af67d3eeccc958f36af210b7e3028db3782016d453cde6766523faaddb467e"} Dec 01 18:27:43 crc kubenswrapper[4868]: I1201 18:27:43.807438 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vddj9/must-gather-k8hct" event={"ID":"1ef4472c-7831-4220-a58e-efb68fa8d78e","Type":"ContainerStarted","Data":"47cdedfebf40b7e4bdee47ce9157f6ee3b7141bf7b4b5d7d372645084f412a1a"} Dec 01 18:27:44 crc kubenswrapper[4868]: I1201 18:27:44.818792 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vddj9/must-gather-k8hct" event={"ID":"1ef4472c-7831-4220-a58e-efb68fa8d78e","Type":"ContainerStarted","Data":"7b5a03b0af4f6f9fe18b4ff51a9da770ebf19f2e548907c2839ee4f51b62ffb0"} Dec 01 18:27:44 crc kubenswrapper[4868]: I1201 18:27:44.841985 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vddj9/must-gather-k8hct" podStartSLOduration=2.841962419 podStartE2EDuration="2.841962419s" podCreationTimestamp="2025-12-01 18:27:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 18:27:44.833027118 +0000 UTC m=+3737.204137539" watchObservedRunningTime="2025-12-01 18:27:44.841962419 +0000 UTC m=+3737.213072830" Dec 01 18:27:47 crc kubenswrapper[4868]: I1201 18:27:47.600593 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vddj9/crc-debug-kgfcb"] Dec 01 18:27:47 crc kubenswrapper[4868]: I1201 18:27:47.602835 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/crc-debug-kgfcb" Dec 01 18:27:47 crc kubenswrapper[4868]: I1201 18:27:47.649575 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkvqz\" (UniqueName: \"kubernetes.io/projected/365325a5-becd-483c-b669-777c7f56f32b-kube-api-access-rkvqz\") pod \"crc-debug-kgfcb\" (UID: \"365325a5-becd-483c-b669-777c7f56f32b\") " pod="openshift-must-gather-vddj9/crc-debug-kgfcb" Dec 01 18:27:47 crc kubenswrapper[4868]: I1201 18:27:47.649698 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/365325a5-becd-483c-b669-777c7f56f32b-host\") pod \"crc-debug-kgfcb\" (UID: \"365325a5-becd-483c-b669-777c7f56f32b\") " pod="openshift-must-gather-vddj9/crc-debug-kgfcb" Dec 01 18:27:47 crc kubenswrapper[4868]: I1201 18:27:47.750975 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkvqz\" (UniqueName: \"kubernetes.io/projected/365325a5-becd-483c-b669-777c7f56f32b-kube-api-access-rkvqz\") pod \"crc-debug-kgfcb\" (UID: \"365325a5-becd-483c-b669-777c7f56f32b\") " pod="openshift-must-gather-vddj9/crc-debug-kgfcb" Dec 01 18:27:47 crc kubenswrapper[4868]: I1201 18:27:47.751104 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/365325a5-becd-483c-b669-777c7f56f32b-host\") pod \"crc-debug-kgfcb\" (UID: \"365325a5-becd-483c-b669-777c7f56f32b\") " pod="openshift-must-gather-vddj9/crc-debug-kgfcb" Dec 01 18:27:47 crc kubenswrapper[4868]: I1201 18:27:47.751263 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/365325a5-becd-483c-b669-777c7f56f32b-host\") pod \"crc-debug-kgfcb\" (UID: \"365325a5-becd-483c-b669-777c7f56f32b\") " pod="openshift-must-gather-vddj9/crc-debug-kgfcb" Dec 01 18:27:47 crc kubenswrapper[4868]: I1201 18:27:47.771504 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkvqz\" (UniqueName: \"kubernetes.io/projected/365325a5-becd-483c-b669-777c7f56f32b-kube-api-access-rkvqz\") pod \"crc-debug-kgfcb\" (UID: \"365325a5-becd-483c-b669-777c7f56f32b\") " pod="openshift-must-gather-vddj9/crc-debug-kgfcb" Dec 01 18:27:47 crc kubenswrapper[4868]: I1201 18:27:47.927824 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/crc-debug-kgfcb" Dec 01 18:27:48 crc kubenswrapper[4868]: I1201 18:27:48.566820 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:48 crc kubenswrapper[4868]: I1201 18:27:48.568301 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:48 crc kubenswrapper[4868]: I1201 18:27:48.622001 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:48 crc kubenswrapper[4868]: I1201 18:27:48.848596 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vddj9/crc-debug-kgfcb" event={"ID":"365325a5-becd-483c-b669-777c7f56f32b","Type":"ContainerStarted","Data":"8660cae0d0a180faf55792f97e80a742d5829b6e1ce21c8d1aa59909e427f544"} Dec 01 18:27:48 crc kubenswrapper[4868]: I1201 18:27:48.848650 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vddj9/crc-debug-kgfcb" event={"ID":"365325a5-becd-483c-b669-777c7f56f32b","Type":"ContainerStarted","Data":"9ddae378aadd6674ad2982087dde360880ffa1a0373b5bd97b22bedaf52367ab"} Dec 01 18:27:48 crc kubenswrapper[4868]: I1201 18:27:48.868319 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vddj9/crc-debug-kgfcb" podStartSLOduration=1.868291175 podStartE2EDuration="1.868291175s" podCreationTimestamp="2025-12-01 18:27:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 18:27:48.860720071 +0000 UTC m=+3741.231830482" watchObservedRunningTime="2025-12-01 18:27:48.868291175 +0000 UTC m=+3741.239401586" Dec 01 18:27:48 crc kubenswrapper[4868]: I1201 18:27:48.906456 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:48 crc kubenswrapper[4868]: I1201 18:27:48.953262 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gd4nc"] Dec 01 18:27:50 crc kubenswrapper[4868]: I1201 18:27:50.878769 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gd4nc" podUID="0bf44ca1-5a2e-4150-b255-f104d6752454" containerName="registry-server" containerID="cri-o://572fec971502506024c948a95e29b602877b68cb2dedd4885c8ab8cd347220ad" gracePeriod=2 Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.407082 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.526015 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctfd4\" (UniqueName: \"kubernetes.io/projected/0bf44ca1-5a2e-4150-b255-f104d6752454-kube-api-access-ctfd4\") pod \"0bf44ca1-5a2e-4150-b255-f104d6752454\" (UID: \"0bf44ca1-5a2e-4150-b255-f104d6752454\") " Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.526052 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bf44ca1-5a2e-4150-b255-f104d6752454-utilities\") pod \"0bf44ca1-5a2e-4150-b255-f104d6752454\" (UID: \"0bf44ca1-5a2e-4150-b255-f104d6752454\") " Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.526099 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bf44ca1-5a2e-4150-b255-f104d6752454-catalog-content\") pod \"0bf44ca1-5a2e-4150-b255-f104d6752454\" (UID: \"0bf44ca1-5a2e-4150-b255-f104d6752454\") " Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.527647 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bf44ca1-5a2e-4150-b255-f104d6752454-utilities" (OuterVolumeSpecName: "utilities") pod "0bf44ca1-5a2e-4150-b255-f104d6752454" (UID: "0bf44ca1-5a2e-4150-b255-f104d6752454"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.534717 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bf44ca1-5a2e-4150-b255-f104d6752454-kube-api-access-ctfd4" (OuterVolumeSpecName: "kube-api-access-ctfd4") pod "0bf44ca1-5a2e-4150-b255-f104d6752454" (UID: "0bf44ca1-5a2e-4150-b255-f104d6752454"). InnerVolumeSpecName "kube-api-access-ctfd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.560060 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bf44ca1-5a2e-4150-b255-f104d6752454-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0bf44ca1-5a2e-4150-b255-f104d6752454" (UID: "0bf44ca1-5a2e-4150-b255-f104d6752454"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.628591 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctfd4\" (UniqueName: \"kubernetes.io/projected/0bf44ca1-5a2e-4150-b255-f104d6752454-kube-api-access-ctfd4\") on node \"crc\" DevicePath \"\"" Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.628624 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bf44ca1-5a2e-4150-b255-f104d6752454-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.628635 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bf44ca1-5a2e-4150-b255-f104d6752454-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.890331 4868 generic.go:334] "Generic (PLEG): container finished" podID="0bf44ca1-5a2e-4150-b255-f104d6752454" containerID="572fec971502506024c948a95e29b602877b68cb2dedd4885c8ab8cd347220ad" exitCode=0 Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.890394 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gd4nc" Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.890390 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gd4nc" event={"ID":"0bf44ca1-5a2e-4150-b255-f104d6752454","Type":"ContainerDied","Data":"572fec971502506024c948a95e29b602877b68cb2dedd4885c8ab8cd347220ad"} Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.890802 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gd4nc" event={"ID":"0bf44ca1-5a2e-4150-b255-f104d6752454","Type":"ContainerDied","Data":"784dff73d038f3565d287a64b2adc7f61161da4e65cf0230f968fbc3113806dc"} Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.890826 4868 scope.go:117] "RemoveContainer" containerID="572fec971502506024c948a95e29b602877b68cb2dedd4885c8ab8cd347220ad" Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.921225 4868 scope.go:117] "RemoveContainer" containerID="349e991ffc7ef1c369dbe6da38e9ea53a6c62948191adbb1f731af7d5d5bd884" Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.930141 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gd4nc"] Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.945865 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gd4nc"] Dec 01 18:27:51 crc kubenswrapper[4868]: I1201 18:27:51.956097 4868 scope.go:117] "RemoveContainer" containerID="a4c05f8b0ccdad667ef20ae3299eceaf676bc7b20d80c81aa429e611edbd3e51" Dec 01 18:27:52 crc kubenswrapper[4868]: I1201 18:27:52.005824 4868 scope.go:117] "RemoveContainer" containerID="572fec971502506024c948a95e29b602877b68cb2dedd4885c8ab8cd347220ad" Dec 01 18:27:52 crc kubenswrapper[4868]: E1201 18:27:52.006424 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"572fec971502506024c948a95e29b602877b68cb2dedd4885c8ab8cd347220ad\": container with ID starting with 572fec971502506024c948a95e29b602877b68cb2dedd4885c8ab8cd347220ad not found: ID does not exist" containerID="572fec971502506024c948a95e29b602877b68cb2dedd4885c8ab8cd347220ad" Dec 01 18:27:52 crc kubenswrapper[4868]: I1201 18:27:52.006485 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"572fec971502506024c948a95e29b602877b68cb2dedd4885c8ab8cd347220ad"} err="failed to get container status \"572fec971502506024c948a95e29b602877b68cb2dedd4885c8ab8cd347220ad\": rpc error: code = NotFound desc = could not find container \"572fec971502506024c948a95e29b602877b68cb2dedd4885c8ab8cd347220ad\": container with ID starting with 572fec971502506024c948a95e29b602877b68cb2dedd4885c8ab8cd347220ad not found: ID does not exist" Dec 01 18:27:52 crc kubenswrapper[4868]: I1201 18:27:52.006521 4868 scope.go:117] "RemoveContainer" containerID="349e991ffc7ef1c369dbe6da38e9ea53a6c62948191adbb1f731af7d5d5bd884" Dec 01 18:27:52 crc kubenswrapper[4868]: E1201 18:27:52.007062 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"349e991ffc7ef1c369dbe6da38e9ea53a6c62948191adbb1f731af7d5d5bd884\": container with ID starting with 349e991ffc7ef1c369dbe6da38e9ea53a6c62948191adbb1f731af7d5d5bd884 not found: ID does not exist" containerID="349e991ffc7ef1c369dbe6da38e9ea53a6c62948191adbb1f731af7d5d5bd884" Dec 01 18:27:52 crc kubenswrapper[4868]: I1201 18:27:52.007160 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"349e991ffc7ef1c369dbe6da38e9ea53a6c62948191adbb1f731af7d5d5bd884"} err="failed to get container status \"349e991ffc7ef1c369dbe6da38e9ea53a6c62948191adbb1f731af7d5d5bd884\": rpc error: code = NotFound desc = could not find container \"349e991ffc7ef1c369dbe6da38e9ea53a6c62948191adbb1f731af7d5d5bd884\": container with ID starting with 349e991ffc7ef1c369dbe6da38e9ea53a6c62948191adbb1f731af7d5d5bd884 not found: ID does not exist" Dec 01 18:27:52 crc kubenswrapper[4868]: I1201 18:27:52.007214 4868 scope.go:117] "RemoveContainer" containerID="a4c05f8b0ccdad667ef20ae3299eceaf676bc7b20d80c81aa429e611edbd3e51" Dec 01 18:27:52 crc kubenswrapper[4868]: E1201 18:27:52.007662 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4c05f8b0ccdad667ef20ae3299eceaf676bc7b20d80c81aa429e611edbd3e51\": container with ID starting with a4c05f8b0ccdad667ef20ae3299eceaf676bc7b20d80c81aa429e611edbd3e51 not found: ID does not exist" containerID="a4c05f8b0ccdad667ef20ae3299eceaf676bc7b20d80c81aa429e611edbd3e51" Dec 01 18:27:52 crc kubenswrapper[4868]: I1201 18:27:52.007702 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4c05f8b0ccdad667ef20ae3299eceaf676bc7b20d80c81aa429e611edbd3e51"} err="failed to get container status \"a4c05f8b0ccdad667ef20ae3299eceaf676bc7b20d80c81aa429e611edbd3e51\": rpc error: code = NotFound desc = could not find container \"a4c05f8b0ccdad667ef20ae3299eceaf676bc7b20d80c81aa429e611edbd3e51\": container with ID starting with a4c05f8b0ccdad667ef20ae3299eceaf676bc7b20d80c81aa429e611edbd3e51 not found: ID does not exist" Dec 01 18:27:52 crc kubenswrapper[4868]: I1201 18:27:52.182882 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bf44ca1-5a2e-4150-b255-f104d6752454" path="/var/lib/kubelet/pods/0bf44ca1-5a2e-4150-b255-f104d6752454/volumes" Dec 01 18:28:22 crc kubenswrapper[4868]: I1201 18:28:22.151539 4868 generic.go:334] "Generic (PLEG): container finished" podID="365325a5-becd-483c-b669-777c7f56f32b" containerID="8660cae0d0a180faf55792f97e80a742d5829b6e1ce21c8d1aa59909e427f544" exitCode=0 Dec 01 18:28:22 crc kubenswrapper[4868]: I1201 18:28:22.151609 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vddj9/crc-debug-kgfcb" event={"ID":"365325a5-becd-483c-b669-777c7f56f32b","Type":"ContainerDied","Data":"8660cae0d0a180faf55792f97e80a742d5829b6e1ce21c8d1aa59909e427f544"} Dec 01 18:28:23 crc kubenswrapper[4868]: I1201 18:28:23.273286 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/crc-debug-kgfcb" Dec 01 18:28:23 crc kubenswrapper[4868]: I1201 18:28:23.315720 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vddj9/crc-debug-kgfcb"] Dec 01 18:28:23 crc kubenswrapper[4868]: I1201 18:28:23.325640 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vddj9/crc-debug-kgfcb"] Dec 01 18:28:23 crc kubenswrapper[4868]: I1201 18:28:23.440633 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkvqz\" (UniqueName: \"kubernetes.io/projected/365325a5-becd-483c-b669-777c7f56f32b-kube-api-access-rkvqz\") pod \"365325a5-becd-483c-b669-777c7f56f32b\" (UID: \"365325a5-becd-483c-b669-777c7f56f32b\") " Dec 01 18:28:23 crc kubenswrapper[4868]: I1201 18:28:23.440699 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/365325a5-becd-483c-b669-777c7f56f32b-host\") pod \"365325a5-becd-483c-b669-777c7f56f32b\" (UID: \"365325a5-becd-483c-b669-777c7f56f32b\") " Dec 01 18:28:23 crc kubenswrapper[4868]: I1201 18:28:23.440887 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/365325a5-becd-483c-b669-777c7f56f32b-host" (OuterVolumeSpecName: "host") pod "365325a5-becd-483c-b669-777c7f56f32b" (UID: "365325a5-becd-483c-b669-777c7f56f32b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 18:28:23 crc kubenswrapper[4868]: I1201 18:28:23.441460 4868 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/365325a5-becd-483c-b669-777c7f56f32b-host\") on node \"crc\" DevicePath \"\"" Dec 01 18:28:23 crc kubenswrapper[4868]: I1201 18:28:23.447338 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/365325a5-becd-483c-b669-777c7f56f32b-kube-api-access-rkvqz" (OuterVolumeSpecName: "kube-api-access-rkvqz") pod "365325a5-becd-483c-b669-777c7f56f32b" (UID: "365325a5-becd-483c-b669-777c7f56f32b"). InnerVolumeSpecName "kube-api-access-rkvqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:28:23 crc kubenswrapper[4868]: I1201 18:28:23.544177 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkvqz\" (UniqueName: \"kubernetes.io/projected/365325a5-becd-483c-b669-777c7f56f32b-kube-api-access-rkvqz\") on node \"crc\" DevicePath \"\"" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.168597 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ddae378aadd6674ad2982087dde360880ffa1a0373b5bd97b22bedaf52367ab" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.168668 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/crc-debug-kgfcb" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.186448 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="365325a5-becd-483c-b669-777c7f56f32b" path="/var/lib/kubelet/pods/365325a5-becd-483c-b669-777c7f56f32b/volumes" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.500822 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vddj9/crc-debug-v4bxp"] Dec 01 18:28:24 crc kubenswrapper[4868]: E1201 18:28:24.501208 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf44ca1-5a2e-4150-b255-f104d6752454" containerName="extract-utilities" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.501220 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf44ca1-5a2e-4150-b255-f104d6752454" containerName="extract-utilities" Dec 01 18:28:24 crc kubenswrapper[4868]: E1201 18:28:24.501243 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="365325a5-becd-483c-b669-777c7f56f32b" containerName="container-00" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.501249 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="365325a5-becd-483c-b669-777c7f56f32b" containerName="container-00" Dec 01 18:28:24 crc kubenswrapper[4868]: E1201 18:28:24.501263 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf44ca1-5a2e-4150-b255-f104d6752454" containerName="extract-content" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.501269 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf44ca1-5a2e-4150-b255-f104d6752454" containerName="extract-content" Dec 01 18:28:24 crc kubenswrapper[4868]: E1201 18:28:24.501281 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf44ca1-5a2e-4150-b255-f104d6752454" containerName="registry-server" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.501287 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf44ca1-5a2e-4150-b255-f104d6752454" containerName="registry-server" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.501472 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="365325a5-becd-483c-b669-777c7f56f32b" containerName="container-00" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.501489 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf44ca1-5a2e-4150-b255-f104d6752454" containerName="registry-server" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.502085 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/crc-debug-v4bxp" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.664515 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7phm\" (UniqueName: \"kubernetes.io/projected/fc250c28-9397-4f7c-bf12-517cdc72072a-kube-api-access-h7phm\") pod \"crc-debug-v4bxp\" (UID: \"fc250c28-9397-4f7c-bf12-517cdc72072a\") " pod="openshift-must-gather-vddj9/crc-debug-v4bxp" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.664892 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc250c28-9397-4f7c-bf12-517cdc72072a-host\") pod \"crc-debug-v4bxp\" (UID: \"fc250c28-9397-4f7c-bf12-517cdc72072a\") " pod="openshift-must-gather-vddj9/crc-debug-v4bxp" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.767222 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7phm\" (UniqueName: \"kubernetes.io/projected/fc250c28-9397-4f7c-bf12-517cdc72072a-kube-api-access-h7phm\") pod \"crc-debug-v4bxp\" (UID: \"fc250c28-9397-4f7c-bf12-517cdc72072a\") " pod="openshift-must-gather-vddj9/crc-debug-v4bxp" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.767347 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc250c28-9397-4f7c-bf12-517cdc72072a-host\") pod \"crc-debug-v4bxp\" (UID: \"fc250c28-9397-4f7c-bf12-517cdc72072a\") " pod="openshift-must-gather-vddj9/crc-debug-v4bxp" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.767624 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc250c28-9397-4f7c-bf12-517cdc72072a-host\") pod \"crc-debug-v4bxp\" (UID: \"fc250c28-9397-4f7c-bf12-517cdc72072a\") " pod="openshift-must-gather-vddj9/crc-debug-v4bxp" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.791738 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7phm\" (UniqueName: \"kubernetes.io/projected/fc250c28-9397-4f7c-bf12-517cdc72072a-kube-api-access-h7phm\") pod \"crc-debug-v4bxp\" (UID: \"fc250c28-9397-4f7c-bf12-517cdc72072a\") " pod="openshift-must-gather-vddj9/crc-debug-v4bxp" Dec 01 18:28:24 crc kubenswrapper[4868]: I1201 18:28:24.824127 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/crc-debug-v4bxp" Dec 01 18:28:25 crc kubenswrapper[4868]: I1201 18:28:25.184163 4868 generic.go:334] "Generic (PLEG): container finished" podID="fc250c28-9397-4f7c-bf12-517cdc72072a" containerID="689cdc795b4d25add302781a0c942577451742c3564b76d796b837d2183f9016" exitCode=0 Dec 01 18:28:25 crc kubenswrapper[4868]: I1201 18:28:25.184393 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vddj9/crc-debug-v4bxp" event={"ID":"fc250c28-9397-4f7c-bf12-517cdc72072a","Type":"ContainerDied","Data":"689cdc795b4d25add302781a0c942577451742c3564b76d796b837d2183f9016"} Dec 01 18:28:25 crc kubenswrapper[4868]: I1201 18:28:25.184467 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vddj9/crc-debug-v4bxp" event={"ID":"fc250c28-9397-4f7c-bf12-517cdc72072a","Type":"ContainerStarted","Data":"a70ee8e479e33915a84ee934b0c9b6640fdfe2636e3c87d34c6a163d4e47671b"} Dec 01 18:28:25 crc kubenswrapper[4868]: I1201 18:28:25.671928 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vddj9/crc-debug-v4bxp"] Dec 01 18:28:25 crc kubenswrapper[4868]: I1201 18:28:25.681957 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vddj9/crc-debug-v4bxp"] Dec 01 18:28:25 crc kubenswrapper[4868]: I1201 18:28:25.904396 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:28:25 crc kubenswrapper[4868]: I1201 18:28:25.904458 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.156466 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vltn2"] Dec 01 18:28:26 crc kubenswrapper[4868]: E1201 18:28:26.157391 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc250c28-9397-4f7c-bf12-517cdc72072a" containerName="container-00" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.157415 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc250c28-9397-4f7c-bf12-517cdc72072a" containerName="container-00" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.157667 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc250c28-9397-4f7c-bf12-517cdc72072a" containerName="container-00" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.159541 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.168687 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vltn2"] Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.300121 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d45423a-808a-421d-ab3d-96baf8eb97f6-catalog-content\") pod \"community-operators-vltn2\" (UID: \"8d45423a-808a-421d-ab3d-96baf8eb97f6\") " pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.300174 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d45423a-808a-421d-ab3d-96baf8eb97f6-utilities\") pod \"community-operators-vltn2\" (UID: \"8d45423a-808a-421d-ab3d-96baf8eb97f6\") " pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.300225 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-667db\" (UniqueName: \"kubernetes.io/projected/8d45423a-808a-421d-ab3d-96baf8eb97f6-kube-api-access-667db\") pod \"community-operators-vltn2\" (UID: \"8d45423a-808a-421d-ab3d-96baf8eb97f6\") " pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.303458 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/crc-debug-v4bxp" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.402468 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc250c28-9397-4f7c-bf12-517cdc72072a-host\") pod \"fc250c28-9397-4f7c-bf12-517cdc72072a\" (UID: \"fc250c28-9397-4f7c-bf12-517cdc72072a\") " Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.402527 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7phm\" (UniqueName: \"kubernetes.io/projected/fc250c28-9397-4f7c-bf12-517cdc72072a-kube-api-access-h7phm\") pod \"fc250c28-9397-4f7c-bf12-517cdc72072a\" (UID: \"fc250c28-9397-4f7c-bf12-517cdc72072a\") " Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.403119 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d45423a-808a-421d-ab3d-96baf8eb97f6-catalog-content\") pod \"community-operators-vltn2\" (UID: \"8d45423a-808a-421d-ab3d-96baf8eb97f6\") " pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.403178 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d45423a-808a-421d-ab3d-96baf8eb97f6-utilities\") pod \"community-operators-vltn2\" (UID: \"8d45423a-808a-421d-ab3d-96baf8eb97f6\") " pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.403248 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-667db\" (UniqueName: \"kubernetes.io/projected/8d45423a-808a-421d-ab3d-96baf8eb97f6-kube-api-access-667db\") pod \"community-operators-vltn2\" (UID: \"8d45423a-808a-421d-ab3d-96baf8eb97f6\") " pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.403622 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc250c28-9397-4f7c-bf12-517cdc72072a-host" (OuterVolumeSpecName: "host") pod "fc250c28-9397-4f7c-bf12-517cdc72072a" (UID: "fc250c28-9397-4f7c-bf12-517cdc72072a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.404567 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d45423a-808a-421d-ab3d-96baf8eb97f6-catalog-content\") pod \"community-operators-vltn2\" (UID: \"8d45423a-808a-421d-ab3d-96baf8eb97f6\") " pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.404591 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d45423a-808a-421d-ab3d-96baf8eb97f6-utilities\") pod \"community-operators-vltn2\" (UID: \"8d45423a-808a-421d-ab3d-96baf8eb97f6\") " pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.412539 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc250c28-9397-4f7c-bf12-517cdc72072a-kube-api-access-h7phm" (OuterVolumeSpecName: "kube-api-access-h7phm") pod "fc250c28-9397-4f7c-bf12-517cdc72072a" (UID: "fc250c28-9397-4f7c-bf12-517cdc72072a"). InnerVolumeSpecName "kube-api-access-h7phm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.419165 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-667db\" (UniqueName: \"kubernetes.io/projected/8d45423a-808a-421d-ab3d-96baf8eb97f6-kube-api-access-667db\") pod \"community-operators-vltn2\" (UID: \"8d45423a-808a-421d-ab3d-96baf8eb97f6\") " pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.501521 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.511069 4868 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc250c28-9397-4f7c-bf12-517cdc72072a-host\") on node \"crc\" DevicePath \"\"" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.511354 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7phm\" (UniqueName: \"kubernetes.io/projected/fc250c28-9397-4f7c-bf12-517cdc72072a-kube-api-access-h7phm\") on node \"crc\" DevicePath \"\"" Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.900852 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vddj9/crc-debug-g4zpb"] Dec 01 18:28:26 crc kubenswrapper[4868]: I1201 18:28:26.903969 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/crc-debug-g4zpb" Dec 01 18:28:27 crc kubenswrapper[4868]: I1201 18:28:27.025223 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b0c03455-423d-430e-96d1-6733cfa8f978-host\") pod \"crc-debug-g4zpb\" (UID: \"b0c03455-423d-430e-96d1-6733cfa8f978\") " pod="openshift-must-gather-vddj9/crc-debug-g4zpb" Dec 01 18:28:27 crc kubenswrapper[4868]: I1201 18:28:27.025265 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj2cs\" (UniqueName: \"kubernetes.io/projected/b0c03455-423d-430e-96d1-6733cfa8f978-kube-api-access-lj2cs\") pod \"crc-debug-g4zpb\" (UID: \"b0c03455-423d-430e-96d1-6733cfa8f978\") " pod="openshift-must-gather-vddj9/crc-debug-g4zpb" Dec 01 18:28:27 crc kubenswrapper[4868]: I1201 18:28:27.061337 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vltn2"] Dec 01 18:28:27 crc kubenswrapper[4868]: W1201 18:28:27.066315 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d45423a_808a_421d_ab3d_96baf8eb97f6.slice/crio-fe4ea803b633a63be34a0fc4d9ca4d08f2bb32745bfeb73312648a41cdea3ccd WatchSource:0}: Error finding container fe4ea803b633a63be34a0fc4d9ca4d08f2bb32745bfeb73312648a41cdea3ccd: Status 404 returned error can't find the container with id fe4ea803b633a63be34a0fc4d9ca4d08f2bb32745bfeb73312648a41cdea3ccd Dec 01 18:28:27 crc kubenswrapper[4868]: I1201 18:28:27.127381 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b0c03455-423d-430e-96d1-6733cfa8f978-host\") pod \"crc-debug-g4zpb\" (UID: \"b0c03455-423d-430e-96d1-6733cfa8f978\") " pod="openshift-must-gather-vddj9/crc-debug-g4zpb" Dec 01 18:28:27 crc kubenswrapper[4868]: I1201 18:28:27.127772 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj2cs\" (UniqueName: \"kubernetes.io/projected/b0c03455-423d-430e-96d1-6733cfa8f978-kube-api-access-lj2cs\") pod \"crc-debug-g4zpb\" (UID: \"b0c03455-423d-430e-96d1-6733cfa8f978\") " pod="openshift-must-gather-vddj9/crc-debug-g4zpb" Dec 01 18:28:27 crc kubenswrapper[4868]: I1201 18:28:27.127538 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b0c03455-423d-430e-96d1-6733cfa8f978-host\") pod \"crc-debug-g4zpb\" (UID: \"b0c03455-423d-430e-96d1-6733cfa8f978\") " pod="openshift-must-gather-vddj9/crc-debug-g4zpb" Dec 01 18:28:27 crc kubenswrapper[4868]: I1201 18:28:27.151318 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj2cs\" (UniqueName: \"kubernetes.io/projected/b0c03455-423d-430e-96d1-6733cfa8f978-kube-api-access-lj2cs\") pod \"crc-debug-g4zpb\" (UID: \"b0c03455-423d-430e-96d1-6733cfa8f978\") " pod="openshift-must-gather-vddj9/crc-debug-g4zpb" Dec 01 18:28:27 crc kubenswrapper[4868]: I1201 18:28:27.227380 4868 scope.go:117] "RemoveContainer" containerID="689cdc795b4d25add302781a0c942577451742c3564b76d796b837d2183f9016" Dec 01 18:28:27 crc kubenswrapper[4868]: I1201 18:28:27.227381 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/crc-debug-v4bxp" Dec 01 18:28:27 crc kubenswrapper[4868]: I1201 18:28:27.228011 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/crc-debug-g4zpb" Dec 01 18:28:27 crc kubenswrapper[4868]: I1201 18:28:27.232855 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vltn2" event={"ID":"8d45423a-808a-421d-ab3d-96baf8eb97f6","Type":"ContainerStarted","Data":"f4d96a5e681161eb60246031d75ebb9d0261397bae37f4b2e732955380021306"} Dec 01 18:28:27 crc kubenswrapper[4868]: I1201 18:28:27.233008 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vltn2" event={"ID":"8d45423a-808a-421d-ab3d-96baf8eb97f6","Type":"ContainerStarted","Data":"fe4ea803b633a63be34a0fc4d9ca4d08f2bb32745bfeb73312648a41cdea3ccd"} Dec 01 18:28:27 crc kubenswrapper[4868]: W1201 18:28:27.264382 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0c03455_423d_430e_96d1_6733cfa8f978.slice/crio-5f94f39d718ff25b5eea48cd967896cc651d944dadf5ce281f78d6b59839c4d1 WatchSource:0}: Error finding container 5f94f39d718ff25b5eea48cd967896cc651d944dadf5ce281f78d6b59839c4d1: Status 404 returned error can't find the container with id 5f94f39d718ff25b5eea48cd967896cc651d944dadf5ce281f78d6b59839c4d1 Dec 01 18:28:28 crc kubenswrapper[4868]: I1201 18:28:28.183202 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc250c28-9397-4f7c-bf12-517cdc72072a" path="/var/lib/kubelet/pods/fc250c28-9397-4f7c-bf12-517cdc72072a/volumes" Dec 01 18:28:28 crc kubenswrapper[4868]: I1201 18:28:28.244457 4868 generic.go:334] "Generic (PLEG): container finished" podID="b0c03455-423d-430e-96d1-6733cfa8f978" containerID="58f2dc73d9a0ed137fe4b560b33ed2820b09c3609deeab5d9c37f748df8b7335" exitCode=0 Dec 01 18:28:28 crc kubenswrapper[4868]: I1201 18:28:28.244535 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vddj9/crc-debug-g4zpb" event={"ID":"b0c03455-423d-430e-96d1-6733cfa8f978","Type":"ContainerDied","Data":"58f2dc73d9a0ed137fe4b560b33ed2820b09c3609deeab5d9c37f748df8b7335"} Dec 01 18:28:28 crc kubenswrapper[4868]: I1201 18:28:28.244592 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vddj9/crc-debug-g4zpb" event={"ID":"b0c03455-423d-430e-96d1-6733cfa8f978","Type":"ContainerStarted","Data":"5f94f39d718ff25b5eea48cd967896cc651d944dadf5ce281f78d6b59839c4d1"} Dec 01 18:28:28 crc kubenswrapper[4868]: I1201 18:28:28.247505 4868 generic.go:334] "Generic (PLEG): container finished" podID="8d45423a-808a-421d-ab3d-96baf8eb97f6" containerID="f4d96a5e681161eb60246031d75ebb9d0261397bae37f4b2e732955380021306" exitCode=0 Dec 01 18:28:28 crc kubenswrapper[4868]: I1201 18:28:28.247543 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vltn2" event={"ID":"8d45423a-808a-421d-ab3d-96baf8eb97f6","Type":"ContainerDied","Data":"f4d96a5e681161eb60246031d75ebb9d0261397bae37f4b2e732955380021306"} Dec 01 18:28:28 crc kubenswrapper[4868]: I1201 18:28:28.289153 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vddj9/crc-debug-g4zpb"] Dec 01 18:28:28 crc kubenswrapper[4868]: I1201 18:28:28.295975 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vddj9/crc-debug-g4zpb"] Dec 01 18:28:29 crc kubenswrapper[4868]: I1201 18:28:29.430496 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/crc-debug-g4zpb" Dec 01 18:28:29 crc kubenswrapper[4868]: I1201 18:28:29.571891 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b0c03455-423d-430e-96d1-6733cfa8f978-host\") pod \"b0c03455-423d-430e-96d1-6733cfa8f978\" (UID: \"b0c03455-423d-430e-96d1-6733cfa8f978\") " Dec 01 18:28:29 crc kubenswrapper[4868]: I1201 18:28:29.572022 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0c03455-423d-430e-96d1-6733cfa8f978-host" (OuterVolumeSpecName: "host") pod "b0c03455-423d-430e-96d1-6733cfa8f978" (UID: "b0c03455-423d-430e-96d1-6733cfa8f978"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 18:28:29 crc kubenswrapper[4868]: I1201 18:28:29.572119 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lj2cs\" (UniqueName: \"kubernetes.io/projected/b0c03455-423d-430e-96d1-6733cfa8f978-kube-api-access-lj2cs\") pod \"b0c03455-423d-430e-96d1-6733cfa8f978\" (UID: \"b0c03455-423d-430e-96d1-6733cfa8f978\") " Dec 01 18:28:29 crc kubenswrapper[4868]: I1201 18:28:29.572499 4868 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b0c03455-423d-430e-96d1-6733cfa8f978-host\") on node \"crc\" DevicePath \"\"" Dec 01 18:28:29 crc kubenswrapper[4868]: I1201 18:28:29.578094 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0c03455-423d-430e-96d1-6733cfa8f978-kube-api-access-lj2cs" (OuterVolumeSpecName: "kube-api-access-lj2cs") pod "b0c03455-423d-430e-96d1-6733cfa8f978" (UID: "b0c03455-423d-430e-96d1-6733cfa8f978"). InnerVolumeSpecName "kube-api-access-lj2cs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:28:29 crc kubenswrapper[4868]: I1201 18:28:29.674260 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lj2cs\" (UniqueName: \"kubernetes.io/projected/b0c03455-423d-430e-96d1-6733cfa8f978-kube-api-access-lj2cs\") on node \"crc\" DevicePath \"\"" Dec 01 18:28:30 crc kubenswrapper[4868]: I1201 18:28:30.182496 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0c03455-423d-430e-96d1-6733cfa8f978" path="/var/lib/kubelet/pods/b0c03455-423d-430e-96d1-6733cfa8f978/volumes" Dec 01 18:28:30 crc kubenswrapper[4868]: I1201 18:28:30.265652 4868 generic.go:334] "Generic (PLEG): container finished" podID="8d45423a-808a-421d-ab3d-96baf8eb97f6" containerID="292160cce1a07b44208d3163cc55016a670fb0053650b2663dcb9f600deedb5f" exitCode=0 Dec 01 18:28:30 crc kubenswrapper[4868]: I1201 18:28:30.265729 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vltn2" event={"ID":"8d45423a-808a-421d-ab3d-96baf8eb97f6","Type":"ContainerDied","Data":"292160cce1a07b44208d3163cc55016a670fb0053650b2663dcb9f600deedb5f"} Dec 01 18:28:30 crc kubenswrapper[4868]: I1201 18:28:30.267853 4868 scope.go:117] "RemoveContainer" containerID="58f2dc73d9a0ed137fe4b560b33ed2820b09c3609deeab5d9c37f748df8b7335" Dec 01 18:28:30 crc kubenswrapper[4868]: I1201 18:28:30.267885 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/crc-debug-g4zpb" Dec 01 18:28:31 crc kubenswrapper[4868]: I1201 18:28:31.277946 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vltn2" event={"ID":"8d45423a-808a-421d-ab3d-96baf8eb97f6","Type":"ContainerStarted","Data":"f8b2546ad11323261ee6d6eac2a2e13aa80cf27e9f2fbe38c42ab151e9716c86"} Dec 01 18:28:31 crc kubenswrapper[4868]: I1201 18:28:31.298463 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vltn2" podStartSLOduration=2.830592653 podStartE2EDuration="5.298443804s" podCreationTimestamp="2025-12-01 18:28:26 +0000 UTC" firstStartedPulling="2025-12-01 18:28:28.249186486 +0000 UTC m=+3780.620296897" lastFinishedPulling="2025-12-01 18:28:30.717037637 +0000 UTC m=+3783.088148048" observedRunningTime="2025-12-01 18:28:31.294721073 +0000 UTC m=+3783.665831504" watchObservedRunningTime="2025-12-01 18:28:31.298443804 +0000 UTC m=+3783.669554215" Dec 01 18:28:36 crc kubenswrapper[4868]: I1201 18:28:36.503055 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:36 crc kubenswrapper[4868]: I1201 18:28:36.503571 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:36 crc kubenswrapper[4868]: I1201 18:28:36.544160 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:37 crc kubenswrapper[4868]: I1201 18:28:37.388571 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:37 crc kubenswrapper[4868]: I1201 18:28:37.433782 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vltn2"] Dec 01 18:28:39 crc kubenswrapper[4868]: I1201 18:28:39.356694 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vltn2" podUID="8d45423a-808a-421d-ab3d-96baf8eb97f6" containerName="registry-server" containerID="cri-o://f8b2546ad11323261ee6d6eac2a2e13aa80cf27e9f2fbe38c42ab151e9716c86" gracePeriod=2 Dec 01 18:28:39 crc kubenswrapper[4868]: I1201 18:28:39.881914 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:39 crc kubenswrapper[4868]: I1201 18:28:39.964741 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d45423a-808a-421d-ab3d-96baf8eb97f6-catalog-content\") pod \"8d45423a-808a-421d-ab3d-96baf8eb97f6\" (UID: \"8d45423a-808a-421d-ab3d-96baf8eb97f6\") " Dec 01 18:28:39 crc kubenswrapper[4868]: I1201 18:28:39.964786 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d45423a-808a-421d-ab3d-96baf8eb97f6-utilities\") pod \"8d45423a-808a-421d-ab3d-96baf8eb97f6\" (UID: \"8d45423a-808a-421d-ab3d-96baf8eb97f6\") " Dec 01 18:28:39 crc kubenswrapper[4868]: I1201 18:28:39.965003 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-667db\" (UniqueName: \"kubernetes.io/projected/8d45423a-808a-421d-ab3d-96baf8eb97f6-kube-api-access-667db\") pod \"8d45423a-808a-421d-ab3d-96baf8eb97f6\" (UID: \"8d45423a-808a-421d-ab3d-96baf8eb97f6\") " Dec 01 18:28:39 crc kubenswrapper[4868]: I1201 18:28:39.966063 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d45423a-808a-421d-ab3d-96baf8eb97f6-utilities" (OuterVolumeSpecName: "utilities") pod "8d45423a-808a-421d-ab3d-96baf8eb97f6" (UID: "8d45423a-808a-421d-ab3d-96baf8eb97f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:28:39 crc kubenswrapper[4868]: I1201 18:28:39.980703 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d45423a-808a-421d-ab3d-96baf8eb97f6-kube-api-access-667db" (OuterVolumeSpecName: "kube-api-access-667db") pod "8d45423a-808a-421d-ab3d-96baf8eb97f6" (UID: "8d45423a-808a-421d-ab3d-96baf8eb97f6"). InnerVolumeSpecName "kube-api-access-667db". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.023328 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d45423a-808a-421d-ab3d-96baf8eb97f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d45423a-808a-421d-ab3d-96baf8eb97f6" (UID: "8d45423a-808a-421d-ab3d-96baf8eb97f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.068567 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-667db\" (UniqueName: \"kubernetes.io/projected/8d45423a-808a-421d-ab3d-96baf8eb97f6-kube-api-access-667db\") on node \"crc\" DevicePath \"\"" Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.068638 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d45423a-808a-421d-ab3d-96baf8eb97f6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.068653 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d45423a-808a-421d-ab3d-96baf8eb97f6-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.369063 4868 generic.go:334] "Generic (PLEG): container finished" podID="8d45423a-808a-421d-ab3d-96baf8eb97f6" containerID="f8b2546ad11323261ee6d6eac2a2e13aa80cf27e9f2fbe38c42ab151e9716c86" exitCode=0 Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.369139 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vltn2" Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.369177 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vltn2" event={"ID":"8d45423a-808a-421d-ab3d-96baf8eb97f6","Type":"ContainerDied","Data":"f8b2546ad11323261ee6d6eac2a2e13aa80cf27e9f2fbe38c42ab151e9716c86"} Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.369527 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vltn2" event={"ID":"8d45423a-808a-421d-ab3d-96baf8eb97f6","Type":"ContainerDied","Data":"fe4ea803b633a63be34a0fc4d9ca4d08f2bb32745bfeb73312648a41cdea3ccd"} Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.369557 4868 scope.go:117] "RemoveContainer" containerID="f8b2546ad11323261ee6d6eac2a2e13aa80cf27e9f2fbe38c42ab151e9716c86" Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.395966 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vltn2"] Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.397873 4868 scope.go:117] "RemoveContainer" containerID="292160cce1a07b44208d3163cc55016a670fb0053650b2663dcb9f600deedb5f" Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.403716 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vltn2"] Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.443729 4868 scope.go:117] "RemoveContainer" containerID="f4d96a5e681161eb60246031d75ebb9d0261397bae37f4b2e732955380021306" Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.467546 4868 scope.go:117] "RemoveContainer" containerID="f8b2546ad11323261ee6d6eac2a2e13aa80cf27e9f2fbe38c42ab151e9716c86" Dec 01 18:28:40 crc kubenswrapper[4868]: E1201 18:28:40.468081 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8b2546ad11323261ee6d6eac2a2e13aa80cf27e9f2fbe38c42ab151e9716c86\": container with ID starting with f8b2546ad11323261ee6d6eac2a2e13aa80cf27e9f2fbe38c42ab151e9716c86 not found: ID does not exist" containerID="f8b2546ad11323261ee6d6eac2a2e13aa80cf27e9f2fbe38c42ab151e9716c86" Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.468128 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8b2546ad11323261ee6d6eac2a2e13aa80cf27e9f2fbe38c42ab151e9716c86"} err="failed to get container status \"f8b2546ad11323261ee6d6eac2a2e13aa80cf27e9f2fbe38c42ab151e9716c86\": rpc error: code = NotFound desc = could not find container \"f8b2546ad11323261ee6d6eac2a2e13aa80cf27e9f2fbe38c42ab151e9716c86\": container with ID starting with f8b2546ad11323261ee6d6eac2a2e13aa80cf27e9f2fbe38c42ab151e9716c86 not found: ID does not exist" Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.468158 4868 scope.go:117] "RemoveContainer" containerID="292160cce1a07b44208d3163cc55016a670fb0053650b2663dcb9f600deedb5f" Dec 01 18:28:40 crc kubenswrapper[4868]: E1201 18:28:40.468682 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"292160cce1a07b44208d3163cc55016a670fb0053650b2663dcb9f600deedb5f\": container with ID starting with 292160cce1a07b44208d3163cc55016a670fb0053650b2663dcb9f600deedb5f not found: ID does not exist" containerID="292160cce1a07b44208d3163cc55016a670fb0053650b2663dcb9f600deedb5f" Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.468728 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"292160cce1a07b44208d3163cc55016a670fb0053650b2663dcb9f600deedb5f"} err="failed to get container status \"292160cce1a07b44208d3163cc55016a670fb0053650b2663dcb9f600deedb5f\": rpc error: code = NotFound desc = could not find container \"292160cce1a07b44208d3163cc55016a670fb0053650b2663dcb9f600deedb5f\": container with ID starting with 292160cce1a07b44208d3163cc55016a670fb0053650b2663dcb9f600deedb5f not found: ID does not exist" Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.468765 4868 scope.go:117] "RemoveContainer" containerID="f4d96a5e681161eb60246031d75ebb9d0261397bae37f4b2e732955380021306" Dec 01 18:28:40 crc kubenswrapper[4868]: E1201 18:28:40.469188 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4d96a5e681161eb60246031d75ebb9d0261397bae37f4b2e732955380021306\": container with ID starting with f4d96a5e681161eb60246031d75ebb9d0261397bae37f4b2e732955380021306 not found: ID does not exist" containerID="f4d96a5e681161eb60246031d75ebb9d0261397bae37f4b2e732955380021306" Dec 01 18:28:40 crc kubenswrapper[4868]: I1201 18:28:40.469220 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4d96a5e681161eb60246031d75ebb9d0261397bae37f4b2e732955380021306"} err="failed to get container status \"f4d96a5e681161eb60246031d75ebb9d0261397bae37f4b2e732955380021306\": rpc error: code = NotFound desc = could not find container \"f4d96a5e681161eb60246031d75ebb9d0261397bae37f4b2e732955380021306\": container with ID starting with f4d96a5e681161eb60246031d75ebb9d0261397bae37f4b2e732955380021306 not found: ID does not exist" Dec 01 18:28:42 crc kubenswrapper[4868]: I1201 18:28:42.183702 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d45423a-808a-421d-ab3d-96baf8eb97f6" path="/var/lib/kubelet/pods/8d45423a-808a-421d-ab3d-96baf8eb97f6/volumes" Dec 01 18:28:50 crc kubenswrapper[4868]: I1201 18:28:50.764026 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-77d99f96cd-fz4zl_610e9c9c-cab0-4843-a120-5dde40363fd2/barbican-api/0.log" Dec 01 18:28:50 crc kubenswrapper[4868]: I1201 18:28:50.963790 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-77d99f96cd-fz4zl_610e9c9c-cab0-4843-a120-5dde40363fd2/barbican-api-log/0.log" Dec 01 18:28:50 crc kubenswrapper[4868]: I1201 18:28:50.988842 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79b8fddb46-xqzzs_0a4cab6a-d367-4eff-aab1-5b20b99f855e/barbican-keystone-listener/0.log" Dec 01 18:28:51 crc kubenswrapper[4868]: I1201 18:28:51.057836 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79b8fddb46-xqzzs_0a4cab6a-d367-4eff-aab1-5b20b99f855e/barbican-keystone-listener-log/0.log" Dec 01 18:28:51 crc kubenswrapper[4868]: I1201 18:28:51.164757 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-655b68fbf5-kxnvg_6fdfb743-319d-4a56-8182-442da79d3ed2/barbican-worker/0.log" Dec 01 18:28:51 crc kubenswrapper[4868]: I1201 18:28:51.192822 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-655b68fbf5-kxnvg_6fdfb743-319d-4a56-8182-442da79d3ed2/barbican-worker-log/0.log" Dec 01 18:28:51 crc kubenswrapper[4868]: I1201 18:28:51.782497 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-bwxzp_af68e45f-ae90-4a6b-a189-7a97b42d0b95/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:28:51 crc kubenswrapper[4868]: I1201 18:28:51.789616 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0831ed8f-c5a0-4273-a32d-4f338a325073/ceilometer-central-agent/0.log" Dec 01 18:28:51 crc kubenswrapper[4868]: I1201 18:28:51.808859 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0831ed8f-c5a0-4273-a32d-4f338a325073/ceilometer-notification-agent/0.log" Dec 01 18:28:51 crc kubenswrapper[4868]: I1201 18:28:51.958139 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0831ed8f-c5a0-4273-a32d-4f338a325073/proxy-httpd/0.log" Dec 01 18:28:51 crc kubenswrapper[4868]: I1201 18:28:51.982538 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0831ed8f-c5a0-4273-a32d-4f338a325073/sg-core/0.log" Dec 01 18:28:52 crc kubenswrapper[4868]: I1201 18:28:52.029247 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_76a7855d-b030-49b6-a58f-4cdfc45d489a/cinder-api/0.log" Dec 01 18:28:52 crc kubenswrapper[4868]: I1201 18:28:52.120786 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_76a7855d-b030-49b6-a58f-4cdfc45d489a/cinder-api-log/0.log" Dec 01 18:28:52 crc kubenswrapper[4868]: I1201 18:28:52.289089 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d8a69c3d-3924-4160-87ec-17a00572a260/cinder-scheduler/0.log" Dec 01 18:28:52 crc kubenswrapper[4868]: I1201 18:28:52.291260 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d8a69c3d-3924-4160-87ec-17a00572a260/probe/0.log" Dec 01 18:28:52 crc kubenswrapper[4868]: I1201 18:28:52.383674 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-rnwwg_fb542a8c-69c4-4a30-9a43-de20584d8e23/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:28:52 crc kubenswrapper[4868]: I1201 18:28:52.497148 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-kt2s8_0d17c1fe-ab49-4510-a2cc-ed17ae06abad/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:28:52 crc kubenswrapper[4868]: I1201 18:28:52.618408 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-4rnqg_7480b802-459b-4029-9bc3-e3d7ba5099de/init/0.log" Dec 01 18:28:52 crc kubenswrapper[4868]: I1201 18:28:52.747361 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-4rnqg_7480b802-459b-4029-9bc3-e3d7ba5099de/init/0.log" Dec 01 18:28:52 crc kubenswrapper[4868]: I1201 18:28:52.787529 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-4rnqg_7480b802-459b-4029-9bc3-e3d7ba5099de/dnsmasq-dns/0.log" Dec 01 18:28:52 crc kubenswrapper[4868]: I1201 18:28:52.824058 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-9wb6b_6e39e56d-e163-4e75-909f-410c72fde805/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:28:53 crc kubenswrapper[4868]: I1201 18:28:53.001860 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4247f042-96f0-4bc5-9ca3-a64ee5616624/glance-httpd/0.log" Dec 01 18:28:53 crc kubenswrapper[4868]: I1201 18:28:53.045084 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4247f042-96f0-4bc5-9ca3-a64ee5616624/glance-log/0.log" Dec 01 18:28:53 crc kubenswrapper[4868]: I1201 18:28:53.158617 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_2bc0d7b8-d497-4339-ad24-c9b75e6e06fc/glance-httpd/0.log" Dec 01 18:28:53 crc kubenswrapper[4868]: I1201 18:28:53.214832 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_2bc0d7b8-d497-4339-ad24-c9b75e6e06fc/glance-log/0.log" Dec 01 18:28:53 crc kubenswrapper[4868]: I1201 18:28:53.365458 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-77567b6ccb-79skk_c50007b0-2158-4940-ac84-c60384469583/horizon/0.log" Dec 01 18:28:53 crc kubenswrapper[4868]: I1201 18:28:53.545322 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-drbph_31db6887-0dfe-44c1-8951-757f2287cd9e/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:28:53 crc kubenswrapper[4868]: I1201 18:28:53.729912 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-77567b6ccb-79skk_c50007b0-2158-4940-ac84-c60384469583/horizon-log/0.log" Dec 01 18:28:53 crc kubenswrapper[4868]: I1201 18:28:53.766468 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-5x8p9_23a40d7f-6024-42d5-9f5f-cbdbc2c8740c/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:28:53 crc kubenswrapper[4868]: I1201 18:28:53.926421 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29410201-xgqzm_42f06b5b-cb48-40da-8a82-f3af58170592/keystone-cron/0.log" Dec 01 18:28:53 crc kubenswrapper[4868]: I1201 18:28:53.952989 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5476cd6644-rgwc7_7366fab1-f630-4998-b189-d1b90a274235/keystone-api/0.log" Dec 01 18:28:54 crc kubenswrapper[4868]: I1201 18:28:54.021557 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_36fb76da-24cc-4d21-898e-107fd20eb6a0/kube-state-metrics/0.log" Dec 01 18:28:54 crc kubenswrapper[4868]: I1201 18:28:54.172416 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-t2sbs_944aa8d9-9aa5-466a-82fd-9768a8b970cb/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:28:54 crc kubenswrapper[4868]: I1201 18:28:54.498421 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5c9f554f77-pjbhr_cb459e0b-9096-4bc1-b553-41156deb1e39/neutron-httpd/0.log" Dec 01 18:28:54 crc kubenswrapper[4868]: I1201 18:28:54.509936 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5c9f554f77-pjbhr_cb459e0b-9096-4bc1-b553-41156deb1e39/neutron-api/0.log" Dec 01 18:28:54 crc kubenswrapper[4868]: I1201 18:28:54.582774 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-v7lx6_6201a05d-611d-4220-b7ae-1b0bf860b4c0/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:28:55 crc kubenswrapper[4868]: I1201 18:28:55.213979 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_47392f20-a05b-44d5-be29-1d7588516c0a/nova-api-log/0.log" Dec 01 18:28:55 crc kubenswrapper[4868]: I1201 18:28:55.229851 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_83a0dfb3-863c-4c80-b46c-e4e3b5fe4098/nova-cell0-conductor-conductor/0.log" Dec 01 18:28:55 crc kubenswrapper[4868]: I1201 18:28:55.515238 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_56ca7cd5-c7ad-43aa-b7ca-2d6f3cac1e7c/nova-cell1-conductor-conductor/0.log" Dec 01 18:28:55 crc kubenswrapper[4868]: I1201 18:28:55.625592 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_4014a0ba-e992-4592-aafd-62e2c90e2ab0/nova-cell1-novncproxy-novncproxy/0.log" Dec 01 18:28:55 crc kubenswrapper[4868]: I1201 18:28:55.651644 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_47392f20-a05b-44d5-be29-1d7588516c0a/nova-api-api/0.log" Dec 01 18:28:55 crc kubenswrapper[4868]: I1201 18:28:55.827911 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-j6jdm_0739d3f5-f361-46f0-872d-23f9d29b7d46/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:28:55 crc kubenswrapper[4868]: I1201 18:28:55.905341 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:28:55 crc kubenswrapper[4868]: I1201 18:28:55.905458 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:28:55 crc kubenswrapper[4868]: I1201 18:28:55.952376 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_76578957-dead-4612-9d4d-bd9d19c589ef/nova-metadata-log/0.log" Dec 01 18:28:56 crc kubenswrapper[4868]: I1201 18:28:56.307268 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_cc48650c-5f3d-4c0a-b570-c7eb8615d504/mysql-bootstrap/0.log" Dec 01 18:28:56 crc kubenswrapper[4868]: I1201 18:28:56.393223 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_ea59e84e-774b-4d8f-af90-1de84fe578d7/nova-scheduler-scheduler/0.log" Dec 01 18:28:56 crc kubenswrapper[4868]: I1201 18:28:56.519774 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_cc48650c-5f3d-4c0a-b570-c7eb8615d504/mysql-bootstrap/0.log" Dec 01 18:28:56 crc kubenswrapper[4868]: I1201 18:28:56.532412 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_cc48650c-5f3d-4c0a-b570-c7eb8615d504/galera/0.log" Dec 01 18:28:56 crc kubenswrapper[4868]: I1201 18:28:56.747126 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_68f3c214-92ad-43c0-b06a-33c6f1f90cfb/mysql-bootstrap/0.log" Dec 01 18:28:56 crc kubenswrapper[4868]: I1201 18:28:56.927340 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_68f3c214-92ad-43c0-b06a-33c6f1f90cfb/mysql-bootstrap/0.log" Dec 01 18:28:56 crc kubenswrapper[4868]: I1201 18:28:56.948031 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_68f3c214-92ad-43c0-b06a-33c6f1f90cfb/galera/0.log" Dec 01 18:28:57 crc kubenswrapper[4868]: I1201 18:28:57.101190 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_3f3f5b1c-3ed5-4d79-b298-474f483b2c4a/openstackclient/0.log" Dec 01 18:28:57 crc kubenswrapper[4868]: I1201 18:28:57.173225 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-87c94_7328948b-0a8c-45d3-91b1-c8317bc12499/openstack-network-exporter/0.log" Dec 01 18:28:57 crc kubenswrapper[4868]: I1201 18:28:57.264886 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_76578957-dead-4612-9d4d-bd9d19c589ef/nova-metadata-metadata/0.log" Dec 01 18:28:57 crc kubenswrapper[4868]: I1201 18:28:57.404115 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rzgnl_8cca4bfb-cddc-425e-abe2-6debea4aa92a/ovsdb-server-init/0.log" Dec 01 18:28:57 crc kubenswrapper[4868]: I1201 18:28:57.616701 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rzgnl_8cca4bfb-cddc-425e-abe2-6debea4aa92a/ovsdb-server-init/0.log" Dec 01 18:28:57 crc kubenswrapper[4868]: I1201 18:28:57.691798 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rzgnl_8cca4bfb-cddc-425e-abe2-6debea4aa92a/ovs-vswitchd/0.log" Dec 01 18:28:57 crc kubenswrapper[4868]: I1201 18:28:57.709205 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rzgnl_8cca4bfb-cddc-425e-abe2-6debea4aa92a/ovsdb-server/0.log" Dec 01 18:28:57 crc kubenswrapper[4868]: I1201 18:28:57.831225 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-z4f4z_65d7a832-930e-4103-90f6-dbc5de8c1ece/ovn-controller/0.log" Dec 01 18:28:57 crc kubenswrapper[4868]: I1201 18:28:57.974677 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-nmn4w_216fe1b0-ec90-4ee7-91e4-aa24476b39b4/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:28:58 crc kubenswrapper[4868]: I1201 18:28:58.046778 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_56ae3649-0322-453b-9088-6807f59d7d96/openstack-network-exporter/0.log" Dec 01 18:28:58 crc kubenswrapper[4868]: I1201 18:28:58.140998 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_56ae3649-0322-453b-9088-6807f59d7d96/ovn-northd/0.log" Dec 01 18:28:58 crc kubenswrapper[4868]: I1201 18:28:58.254896 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_faa33685-51f4-46ce-bdc7-ef9a2467f482/ovsdbserver-nb/0.log" Dec 01 18:28:58 crc kubenswrapper[4868]: I1201 18:28:58.294749 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_faa33685-51f4-46ce-bdc7-ef9a2467f482/openstack-network-exporter/0.log" Dec 01 18:28:58 crc kubenswrapper[4868]: I1201 18:28:58.500788 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_89b376d8-f72b-4c03-a2f1-3fcf6ff20336/openstack-network-exporter/0.log" Dec 01 18:28:58 crc kubenswrapper[4868]: I1201 18:28:58.535044 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_89b376d8-f72b-4c03-a2f1-3fcf6ff20336/ovsdbserver-sb/0.log" Dec 01 18:28:58 crc kubenswrapper[4868]: I1201 18:28:58.716061 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d8466bd7b-sr5pl_f41cf3a5-eb10-42c0-93a1-f1195331f807/placement-api/0.log" Dec 01 18:28:58 crc kubenswrapper[4868]: I1201 18:28:58.763778 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d8466bd7b-sr5pl_f41cf3a5-eb10-42c0-93a1-f1195331f807/placement-log/0.log" Dec 01 18:28:58 crc kubenswrapper[4868]: I1201 18:28:58.827451 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_12c9fc00-91a0-493e-b6b3-5d1a5ba7685c/setup-container/0.log" Dec 01 18:28:59 crc kubenswrapper[4868]: I1201 18:28:59.053198 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_12c9fc00-91a0-493e-b6b3-5d1a5ba7685c/rabbitmq/0.log" Dec 01 18:28:59 crc kubenswrapper[4868]: I1201 18:28:59.058288 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_12c9fc00-91a0-493e-b6b3-5d1a5ba7685c/setup-container/0.log" Dec 01 18:28:59 crc kubenswrapper[4868]: I1201 18:28:59.068651 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b05cb910-a007-46d5-8182-eb1d30c5f953/setup-container/0.log" Dec 01 18:28:59 crc kubenswrapper[4868]: I1201 18:28:59.319578 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b05cb910-a007-46d5-8182-eb1d30c5f953/setup-container/0.log" Dec 01 18:28:59 crc kubenswrapper[4868]: I1201 18:28:59.335259 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b05cb910-a007-46d5-8182-eb1d30c5f953/rabbitmq/0.log" Dec 01 18:28:59 crc kubenswrapper[4868]: I1201 18:28:59.338917 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-zzqq9_17fd50a0-00e4-4b51-9631-30b48bd3013f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:28:59 crc kubenswrapper[4868]: I1201 18:28:59.499522 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-vvbkm_e989defc-ac68-4417-9bf3-f5962bfdb601/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:28:59 crc kubenswrapper[4868]: I1201 18:28:59.634204 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-bhnfb_f36695d5-3d20-4e3a-9ef3-63cc50a3445c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:28:59 crc kubenswrapper[4868]: I1201 18:28:59.805355 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-q9z9p_2163d223-acf7-452f-8a2e-5461307cf82e/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:28:59 crc kubenswrapper[4868]: I1201 18:28:59.872429 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-zjnp8_3f7f2561-e242-48b1-a363-50bdecab86de/ssh-known-hosts-edpm-deployment/0.log" Dec 01 18:29:00 crc kubenswrapper[4868]: I1201 18:29:00.087231 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-86984fb57-t5hzr_a427adaa-2060-4639-bd6d-e53b8fb00357/proxy-server/0.log" Dec 01 18:29:00 crc kubenswrapper[4868]: I1201 18:29:00.195089 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-86984fb57-t5hzr_a427adaa-2060-4639-bd6d-e53b8fb00357/proxy-httpd/0.log" Dec 01 18:29:00 crc kubenswrapper[4868]: I1201 18:29:00.210164 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-bfdvt_71e5e846-0b2a-47d3-b60b-daf23c22af52/swift-ring-rebalance/0.log" Dec 01 18:29:00 crc kubenswrapper[4868]: I1201 18:29:00.345363 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/account-auditor/0.log" Dec 01 18:29:00 crc kubenswrapper[4868]: I1201 18:29:00.361965 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/account-reaper/0.log" Dec 01 18:29:00 crc kubenswrapper[4868]: I1201 18:29:00.457504 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/account-replicator/0.log" Dec 01 18:29:00 crc kubenswrapper[4868]: I1201 18:29:00.526917 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/account-server/0.log" Dec 01 18:29:00 crc kubenswrapper[4868]: I1201 18:29:00.551636 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/container-auditor/0.log" Dec 01 18:29:00 crc kubenswrapper[4868]: I1201 18:29:00.597981 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/container-replicator/0.log" Dec 01 18:29:00 crc kubenswrapper[4868]: I1201 18:29:00.696743 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/container-server/0.log" Dec 01 18:29:00 crc kubenswrapper[4868]: I1201 18:29:00.756949 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/object-auditor/0.log" Dec 01 18:29:00 crc kubenswrapper[4868]: I1201 18:29:00.803753 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/container-updater/0.log" Dec 01 18:29:00 crc kubenswrapper[4868]: I1201 18:29:00.893193 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/object-expirer/0.log" Dec 01 18:29:00 crc kubenswrapper[4868]: I1201 18:29:00.898628 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/object-replicator/0.log" Dec 01 18:29:00 crc kubenswrapper[4868]: I1201 18:29:00.942608 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/object-server/0.log" Dec 01 18:29:01 crc kubenswrapper[4868]: I1201 18:29:01.038857 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/object-updater/0.log" Dec 01 18:29:01 crc kubenswrapper[4868]: I1201 18:29:01.127285 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/rsync/0.log" Dec 01 18:29:01 crc kubenswrapper[4868]: I1201 18:29:01.159286 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4997c1c8-2ad2-4b4e-b2ff-6980fcb743e4/swift-recon-cron/0.log" Dec 01 18:29:01 crc kubenswrapper[4868]: I1201 18:29:01.327440 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-d6lmq_64998ba8-012c-4ffc-a7b3-dbbbb41940c0/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:29:01 crc kubenswrapper[4868]: I1201 18:29:01.371277 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_8447d0d7-b9fd-4a48-bf54-e43272539aec/tempest-tests-tempest-tests-runner/0.log" Dec 01 18:29:01 crc kubenswrapper[4868]: I1201 18:29:01.582191 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_66487491-7ab1-4ea0-b4df-e7be8a20a7e5/test-operator-logs-container/0.log" Dec 01 18:29:01 crc kubenswrapper[4868]: I1201 18:29:01.650092 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-v6t68_f1d81003-13ab-4833-9b01-3b4c6f8876ef/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 18:29:11 crc kubenswrapper[4868]: I1201 18:29:11.659601 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_4d1ee35e-4e53-46a0-a1d9-86eeb1f9c15e/memcached/0.log" Dec 01 18:29:25 crc kubenswrapper[4868]: I1201 18:29:25.904326 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:29:25 crc kubenswrapper[4868]: I1201 18:29:25.904818 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 18:29:25 crc kubenswrapper[4868]: I1201 18:29:25.904870 4868 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" Dec 01 18:29:25 crc kubenswrapper[4868]: I1201 18:29:25.905601 4868 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f"} pod="openshift-machine-config-operator/machine-config-daemon-njgpn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 18:29:25 crc kubenswrapper[4868]: I1201 18:29:25.905651 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" containerID="cri-o://91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" gracePeriod=600 Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.002152 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts_127f8203-d151-4430-97ae-405c8788a2af/util/0.log" Dec 01 18:29:26 crc kubenswrapper[4868]: E1201 18:29:26.083054 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.233150 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts_127f8203-d151-4430-97ae-405c8788a2af/pull/0.log" Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.246741 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts_127f8203-d151-4430-97ae-405c8788a2af/util/0.log" Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.279907 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts_127f8203-d151-4430-97ae-405c8788a2af/pull/0.log" Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.407734 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts_127f8203-d151-4430-97ae-405c8788a2af/util/0.log" Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.428853 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts_127f8203-d151-4430-97ae-405c8788a2af/pull/0.log" Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.439383 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_3db9e85e220f6e10649f12f77a953e44861605b506a150877b56b0b0d9g87ts_127f8203-d151-4430-97ae-405c8788a2af/extract/0.log" Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.593761 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-8m6s9_f3eeed9b-dfbf-4e17-b80e-8792d471f1fb/kube-rbac-proxy/0.log" Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.676561 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-8m6s9_f3eeed9b-dfbf-4e17-b80e-8792d471f1fb/manager/0.log" Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.693112 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-pk4dk_8cc7f247-f40c-4486-bd86-5f49036a4b2e/kube-rbac-proxy/0.log" Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.796932 4868 generic.go:334] "Generic (PLEG): container finished" podID="ec4e08bf-df56-4490-8e79-a4175c727405" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" exitCode=0 Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.796989 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerDied","Data":"91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f"} Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.797389 4868 scope.go:117] "RemoveContainer" containerID="d96440aad2ff00174b44608542db90d83651ffced39d42cc58197244d3a95475" Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.798191 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:29:26 crc kubenswrapper[4868]: E1201 18:29:26.798473 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.863100 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-pk4dk_8cc7f247-f40c-4486-bd86-5f49036a4b2e/manager/0.log" Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.864310 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-4s44v_a11d06ed-a89a-4600-8c63-4a3845eb01e6/kube-rbac-proxy/0.log" Dec 01 18:29:26 crc kubenswrapper[4868]: I1201 18:29:26.911000 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-4s44v_a11d06ed-a89a-4600-8c63-4a3845eb01e6/manager/0.log" Dec 01 18:29:27 crc kubenswrapper[4868]: I1201 18:29:27.013468 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-rhj2q_6faacf07-6354-4314-829b-db239c85d98e/kube-rbac-proxy/0.log" Dec 01 18:29:27 crc kubenswrapper[4868]: I1201 18:29:27.151311 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-rhj2q_6faacf07-6354-4314-829b-db239c85d98e/manager/0.log" Dec 01 18:29:27 crc kubenswrapper[4868]: I1201 18:29:27.555798 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-kbb5v_c5c836e6-8b19-44f8-91ae-4893dcbfd0fc/kube-rbac-proxy/0.log" Dec 01 18:29:27 crc kubenswrapper[4868]: I1201 18:29:27.591072 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-kbb5v_c5c836e6-8b19-44f8-91ae-4893dcbfd0fc/manager/0.log" Dec 01 18:29:27 crc kubenswrapper[4868]: I1201 18:29:27.756415 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5kkb2_93fc4a4e-c6fb-4d43-88dd-0c24bcd85c5e/kube-rbac-proxy/0.log" Dec 01 18:29:27 crc kubenswrapper[4868]: I1201 18:29:27.786340 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5kkb2_93fc4a4e-c6fb-4d43-88dd-0c24bcd85c5e/manager/0.log" Dec 01 18:29:27 crc kubenswrapper[4868]: I1201 18:29:27.805208 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-55599fd5c4-2vg29_775bad96-f2da-4628-8242-af4778e37aee/kube-rbac-proxy/0.log" Dec 01 18:29:28 crc kubenswrapper[4868]: I1201 18:29:28.046783 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-c9nqt_5cad124c-35fd-404e-bcf4-3b17e89e1288/kube-rbac-proxy/0.log" Dec 01 18:29:28 crc kubenswrapper[4868]: I1201 18:29:28.050985 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-c9nqt_5cad124c-35fd-404e-bcf4-3b17e89e1288/manager/0.log" Dec 01 18:29:28 crc kubenswrapper[4868]: I1201 18:29:28.067688 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-55599fd5c4-2vg29_775bad96-f2da-4628-8242-af4778e37aee/manager/0.log" Dec 01 18:29:28 crc kubenswrapper[4868]: I1201 18:29:28.249336 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-hqmbs_ccf55d21-7eac-4a3e-90c6-147287c4aac3/kube-rbac-proxy/0.log" Dec 01 18:29:28 crc kubenswrapper[4868]: I1201 18:29:28.360765 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-hqmbs_ccf55d21-7eac-4a3e-90c6-147287c4aac3/manager/0.log" Dec 01 18:29:28 crc kubenswrapper[4868]: I1201 18:29:28.510491 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-jjvvf_592806bc-2a42-4c69-b36f-1b8dc3c14249/kube-rbac-proxy/0.log" Dec 01 18:29:28 crc kubenswrapper[4868]: I1201 18:29:28.538513 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-jjvvf_592806bc-2a42-4c69-b36f-1b8dc3c14249/manager/0.log" Dec 01 18:29:28 crc kubenswrapper[4868]: I1201 18:29:28.589410 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-9j2nw_4e7cc027-32a6-437e-b05b-52c2984e0a61/kube-rbac-proxy/0.log" Dec 01 18:29:28 crc kubenswrapper[4868]: I1201 18:29:28.716920 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-9j2nw_4e7cc027-32a6-437e-b05b-52c2984e0a61/manager/0.log" Dec 01 18:29:28 crc kubenswrapper[4868]: I1201 18:29:28.846224 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-7mptb_97f09844-bdf4-4612-b4b4-966ae8dd49c2/kube-rbac-proxy/0.log" Dec 01 18:29:28 crc kubenswrapper[4868]: I1201 18:29:28.883759 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-7mptb_97f09844-bdf4-4612-b4b4-966ae8dd49c2/manager/0.log" Dec 01 18:29:29 crc kubenswrapper[4868]: I1201 18:29:29.023957 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-f7k69_b8df3e8b-0fc5-465f-b531-474518a1b809/kube-rbac-proxy/0.log" Dec 01 18:29:29 crc kubenswrapper[4868]: I1201 18:29:29.163756 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-fmkmw_20c07d22-1628-40a4-97aa-605a2da611df/kube-rbac-proxy/0.log" Dec 01 18:29:29 crc kubenswrapper[4868]: I1201 18:29:29.169587 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-f7k69_b8df3e8b-0fc5-465f-b531-474518a1b809/manager/0.log" Dec 01 18:29:29 crc kubenswrapper[4868]: I1201 18:29:29.259355 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-fmkmw_20c07d22-1628-40a4-97aa-605a2da611df/manager/0.log" Dec 01 18:29:30 crc kubenswrapper[4868]: I1201 18:29:30.045263 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs_25f93b71-f929-4f74-a2d9-822f0b9402bb/kube-rbac-proxy/0.log" Dec 01 18:29:30 crc kubenswrapper[4868]: I1201 18:29:30.056003 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4plgxs_25f93b71-f929-4f74-a2d9-822f0b9402bb/manager/0.log" Dec 01 18:29:30 crc kubenswrapper[4868]: I1201 18:29:30.395796 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-8rf2p_d73d228d-a5b1-4569-8c1f-a062812250b9/registry-server/0.log" Dec 01 18:29:30 crc kubenswrapper[4868]: I1201 18:29:30.536993 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7b559cfbd8-zw5pm_b6698ef2-c594-4859-ba6d-7d76127c819d/operator/0.log" Dec 01 18:29:30 crc kubenswrapper[4868]: I1201 18:29:30.554519 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-97g6w_d87f1b79-a4fb-4cfa-86d7-823ceedcf913/kube-rbac-proxy/0.log" Dec 01 18:29:30 crc kubenswrapper[4868]: I1201 18:29:30.681033 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-97g6w_d87f1b79-a4fb-4cfa-86d7-823ceedcf913/manager/0.log" Dec 01 18:29:30 crc kubenswrapper[4868]: I1201 18:29:30.868319 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-mmjrx_d5d35af7-2b67-42cd-ae7c-d73f105da292/kube-rbac-proxy/0.log" Dec 01 18:29:30 crc kubenswrapper[4868]: I1201 18:29:30.878934 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-mmjrx_d5d35af7-2b67-42cd-ae7c-d73f105da292/manager/0.log" Dec 01 18:29:31 crc kubenswrapper[4868]: I1201 18:29:31.052244 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-blw9f_32e03cbc-4f29-49ee-8558-3f8950b0a383/operator/0.log" Dec 01 18:29:31 crc kubenswrapper[4868]: I1201 18:29:31.115841 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-g56bv_58e2eaed-6851-4568-a604-397a8944da7f/kube-rbac-proxy/0.log" Dec 01 18:29:31 crc kubenswrapper[4868]: I1201 18:29:31.147761 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-g56bv_58e2eaed-6851-4568-a604-397a8944da7f/manager/0.log" Dec 01 18:29:31 crc kubenswrapper[4868]: I1201 18:29:31.316777 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-l4hqj_8df8451f-54fc-4553-834b-37839d4c1807/kube-rbac-proxy/0.log" Dec 01 18:29:31 crc kubenswrapper[4868]: I1201 18:29:31.378760 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-755b486884-bvzjt_515c8a14-67ad-4c4c-8b3e-9433997fbdb3/manager/0.log" Dec 01 18:29:31 crc kubenswrapper[4868]: I1201 18:29:31.479466 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-l4hqj_8df8451f-54fc-4553-834b-37839d4c1807/manager/0.log" Dec 01 18:29:31 crc kubenswrapper[4868]: I1201 18:29:31.494275 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-d2jx2_f1f47ec9-ea94-4fd5-b113-baa29c640369/kube-rbac-proxy/0.log" Dec 01 18:29:31 crc kubenswrapper[4868]: I1201 18:29:31.580194 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-d2jx2_f1f47ec9-ea94-4fd5-b113-baa29c640369/manager/0.log" Dec 01 18:29:31 crc kubenswrapper[4868]: I1201 18:29:31.652715 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-b78w5_fc4efb42-45ab-4341-b6f2-d95497223de3/kube-rbac-proxy/0.log" Dec 01 18:29:31 crc kubenswrapper[4868]: I1201 18:29:31.713809 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-b78w5_fc4efb42-45ab-4341-b6f2-d95497223de3/manager/0.log" Dec 01 18:29:38 crc kubenswrapper[4868]: I1201 18:29:38.181668 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:29:38 crc kubenswrapper[4868]: E1201 18:29:38.182598 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:29:49 crc kubenswrapper[4868]: I1201 18:29:49.150961 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-vw9z9_2bf15be5-1380-4b58-934d-7793e22541b4/control-plane-machine-set-operator/0.log" Dec 01 18:29:49 crc kubenswrapper[4868]: I1201 18:29:49.329603 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qw86w_be098182-dd51-4a19-920a-a797a175124a/kube-rbac-proxy/0.log" Dec 01 18:29:49 crc kubenswrapper[4868]: I1201 18:29:49.362145 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qw86w_be098182-dd51-4a19-920a-a797a175124a/machine-api-operator/0.log" Dec 01 18:29:50 crc kubenswrapper[4868]: I1201 18:29:50.172314 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:29:50 crc kubenswrapper[4868]: E1201 18:29:50.172537 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.183153 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf"] Dec 01 18:30:00 crc kubenswrapper[4868]: E1201 18:30:00.184735 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d45423a-808a-421d-ab3d-96baf8eb97f6" containerName="extract-content" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.184753 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d45423a-808a-421d-ab3d-96baf8eb97f6" containerName="extract-content" Dec 01 18:30:00 crc kubenswrapper[4868]: E1201 18:30:00.184779 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d45423a-808a-421d-ab3d-96baf8eb97f6" containerName="registry-server" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.184789 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d45423a-808a-421d-ab3d-96baf8eb97f6" containerName="registry-server" Dec 01 18:30:00 crc kubenswrapper[4868]: E1201 18:30:00.184802 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d45423a-808a-421d-ab3d-96baf8eb97f6" containerName="extract-utilities" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.184809 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d45423a-808a-421d-ab3d-96baf8eb97f6" containerName="extract-utilities" Dec 01 18:30:00 crc kubenswrapper[4868]: E1201 18:30:00.184842 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0c03455-423d-430e-96d1-6733cfa8f978" containerName="container-00" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.184855 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0c03455-423d-430e-96d1-6733cfa8f978" containerName="container-00" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.185211 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d45423a-808a-421d-ab3d-96baf8eb97f6" containerName="registry-server" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.185224 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0c03455-423d-430e-96d1-6733cfa8f978" containerName="container-00" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.186251 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.186500 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf"] Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.188981 4868 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.190660 4868 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.226738 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/099c1666-e6f3-4192-b083-b2400e6e6c68-secret-volume\") pod \"collect-profiles-29410230-b6gzf\" (UID: \"099c1666-e6f3-4192-b083-b2400e6e6c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.226888 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/099c1666-e6f3-4192-b083-b2400e6e6c68-config-volume\") pod \"collect-profiles-29410230-b6gzf\" (UID: \"099c1666-e6f3-4192-b083-b2400e6e6c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.226996 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6nrx\" (UniqueName: \"kubernetes.io/projected/099c1666-e6f3-4192-b083-b2400e6e6c68-kube-api-access-v6nrx\") pod \"collect-profiles-29410230-b6gzf\" (UID: \"099c1666-e6f3-4192-b083-b2400e6e6c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.329892 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/099c1666-e6f3-4192-b083-b2400e6e6c68-config-volume\") pod \"collect-profiles-29410230-b6gzf\" (UID: \"099c1666-e6f3-4192-b083-b2400e6e6c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.330422 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6nrx\" (UniqueName: \"kubernetes.io/projected/099c1666-e6f3-4192-b083-b2400e6e6c68-kube-api-access-v6nrx\") pod \"collect-profiles-29410230-b6gzf\" (UID: \"099c1666-e6f3-4192-b083-b2400e6e6c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.330577 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/099c1666-e6f3-4192-b083-b2400e6e6c68-secret-volume\") pod \"collect-profiles-29410230-b6gzf\" (UID: \"099c1666-e6f3-4192-b083-b2400e6e6c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.331986 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/099c1666-e6f3-4192-b083-b2400e6e6c68-config-volume\") pod \"collect-profiles-29410230-b6gzf\" (UID: \"099c1666-e6f3-4192-b083-b2400e6e6c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.340033 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/099c1666-e6f3-4192-b083-b2400e6e6c68-secret-volume\") pod \"collect-profiles-29410230-b6gzf\" (UID: \"099c1666-e6f3-4192-b083-b2400e6e6c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.351006 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6nrx\" (UniqueName: \"kubernetes.io/projected/099c1666-e6f3-4192-b083-b2400e6e6c68-kube-api-access-v6nrx\") pod \"collect-profiles-29410230-b6gzf\" (UID: \"099c1666-e6f3-4192-b083-b2400e6e6c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.508521 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.643713 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-f6tcg_958205b1-cd35-41d8-9e21-b30cc80a2489/cert-manager-controller/0.log" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.861638 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-j2dnr_8057b33e-2417-43f9-aab1-7f7db221d051/cert-manager-cainjector/0.log" Dec 01 18:30:00 crc kubenswrapper[4868]: I1201 18:30:00.977350 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-qgmqm_3858baab-f6b9-42a5-b7ad-2c93a8e98fb6/cert-manager-webhook/0.log" Dec 01 18:30:01 crc kubenswrapper[4868]: I1201 18:30:01.064826 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf"] Dec 01 18:30:01 crc kubenswrapper[4868]: I1201 18:30:01.111252 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" event={"ID":"099c1666-e6f3-4192-b083-b2400e6e6c68","Type":"ContainerStarted","Data":"d1a9a5a662764163aa17c1907140eea508bfec0305b4e032c095b0939e8b9715"} Dec 01 18:30:02 crc kubenswrapper[4868]: I1201 18:30:02.123377 4868 generic.go:334] "Generic (PLEG): container finished" podID="099c1666-e6f3-4192-b083-b2400e6e6c68" containerID="e1605399542b61f47cdd9ed47a8e75df2f19cd2206840de070a8aa26fba3c590" exitCode=0 Dec 01 18:30:02 crc kubenswrapper[4868]: I1201 18:30:02.123425 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" event={"ID":"099c1666-e6f3-4192-b083-b2400e6e6c68","Type":"ContainerDied","Data":"e1605399542b61f47cdd9ed47a8e75df2f19cd2206840de070a8aa26fba3c590"} Dec 01 18:30:03 crc kubenswrapper[4868]: I1201 18:30:03.491233 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" Dec 01 18:30:03 crc kubenswrapper[4868]: I1201 18:30:03.641405 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6nrx\" (UniqueName: \"kubernetes.io/projected/099c1666-e6f3-4192-b083-b2400e6e6c68-kube-api-access-v6nrx\") pod \"099c1666-e6f3-4192-b083-b2400e6e6c68\" (UID: \"099c1666-e6f3-4192-b083-b2400e6e6c68\") " Dec 01 18:30:03 crc kubenswrapper[4868]: I1201 18:30:03.641717 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/099c1666-e6f3-4192-b083-b2400e6e6c68-config-volume\") pod \"099c1666-e6f3-4192-b083-b2400e6e6c68\" (UID: \"099c1666-e6f3-4192-b083-b2400e6e6c68\") " Dec 01 18:30:03 crc kubenswrapper[4868]: I1201 18:30:03.641776 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/099c1666-e6f3-4192-b083-b2400e6e6c68-secret-volume\") pod \"099c1666-e6f3-4192-b083-b2400e6e6c68\" (UID: \"099c1666-e6f3-4192-b083-b2400e6e6c68\") " Dec 01 18:30:03 crc kubenswrapper[4868]: I1201 18:30:03.642341 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/099c1666-e6f3-4192-b083-b2400e6e6c68-config-volume" (OuterVolumeSpecName: "config-volume") pod "099c1666-e6f3-4192-b083-b2400e6e6c68" (UID: "099c1666-e6f3-4192-b083-b2400e6e6c68"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 18:30:03 crc kubenswrapper[4868]: I1201 18:30:03.647995 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/099c1666-e6f3-4192-b083-b2400e6e6c68-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "099c1666-e6f3-4192-b083-b2400e6e6c68" (UID: "099c1666-e6f3-4192-b083-b2400e6e6c68"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 18:30:03 crc kubenswrapper[4868]: I1201 18:30:03.650212 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/099c1666-e6f3-4192-b083-b2400e6e6c68-kube-api-access-v6nrx" (OuterVolumeSpecName: "kube-api-access-v6nrx") pod "099c1666-e6f3-4192-b083-b2400e6e6c68" (UID: "099c1666-e6f3-4192-b083-b2400e6e6c68"). InnerVolumeSpecName "kube-api-access-v6nrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:30:03 crc kubenswrapper[4868]: I1201 18:30:03.744431 4868 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/099c1666-e6f3-4192-b083-b2400e6e6c68-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 18:30:03 crc kubenswrapper[4868]: I1201 18:30:03.744466 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6nrx\" (UniqueName: \"kubernetes.io/projected/099c1666-e6f3-4192-b083-b2400e6e6c68-kube-api-access-v6nrx\") on node \"crc\" DevicePath \"\"" Dec 01 18:30:03 crc kubenswrapper[4868]: I1201 18:30:03.744476 4868 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/099c1666-e6f3-4192-b083-b2400e6e6c68-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 18:30:04 crc kubenswrapper[4868]: I1201 18:30:04.141226 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" event={"ID":"099c1666-e6f3-4192-b083-b2400e6e6c68","Type":"ContainerDied","Data":"d1a9a5a662764163aa17c1907140eea508bfec0305b4e032c095b0939e8b9715"} Dec 01 18:30:04 crc kubenswrapper[4868]: I1201 18:30:04.141532 4868 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1a9a5a662764163aa17c1907140eea508bfec0305b4e032c095b0939e8b9715" Dec 01 18:30:04 crc kubenswrapper[4868]: I1201 18:30:04.141302 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410230-b6gzf" Dec 01 18:30:04 crc kubenswrapper[4868]: I1201 18:30:04.172267 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:30:04 crc kubenswrapper[4868]: E1201 18:30:04.172763 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:30:04 crc kubenswrapper[4868]: I1201 18:30:04.578600 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls"] Dec 01 18:30:04 crc kubenswrapper[4868]: I1201 18:30:04.589214 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410185-2m6ls"] Dec 01 18:30:06 crc kubenswrapper[4868]: I1201 18:30:06.181973 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e91d120-4deb-4bc5-a4ae-b277d85ad10a" path="/var/lib/kubelet/pods/7e91d120-4deb-4bc5-a4ae-b277d85ad10a/volumes" Dec 01 18:30:13 crc kubenswrapper[4868]: I1201 18:30:13.019791 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-xsj4b_fa121055-74c0-44aa-8401-3a8476c7db4c/nmstate-console-plugin/0.log" Dec 01 18:30:13 crc kubenswrapper[4868]: I1201 18:30:13.168331 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-67rpb_d04c81f8-7543-44b9-abc9-b0c95346cd56/nmstate-handler/0.log" Dec 01 18:30:13 crc kubenswrapper[4868]: I1201 18:30:13.199999 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-j7lcc_ed23de21-61e1-463e-9a8e-5a8ba51cc727/kube-rbac-proxy/0.log" Dec 01 18:30:13 crc kubenswrapper[4868]: I1201 18:30:13.236555 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-j7lcc_ed23de21-61e1-463e-9a8e-5a8ba51cc727/nmstate-metrics/0.log" Dec 01 18:30:13 crc kubenswrapper[4868]: I1201 18:30:13.398550 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-rb8wl_13c419e8-4943-4ee5-8d1c-5dd5051c81db/nmstate-operator/0.log" Dec 01 18:30:13 crc kubenswrapper[4868]: I1201 18:30:13.411692 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-kjwkw_f1784373-a1b0-4ca1-8c08-cf01fbc83b72/nmstate-webhook/0.log" Dec 01 18:30:15 crc kubenswrapper[4868]: I1201 18:30:15.172579 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:30:15 crc kubenswrapper[4868]: E1201 18:30:15.173111 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:30:26 crc kubenswrapper[4868]: I1201 18:30:26.172471 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:30:26 crc kubenswrapper[4868]: E1201 18:30:26.173260 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:30:27 crc kubenswrapper[4868]: I1201 18:30:27.088003 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-nvnm5_5536b1ca-ecee-4f69-9884-539b000d00f0/kube-rbac-proxy/0.log" Dec 01 18:30:27 crc kubenswrapper[4868]: I1201 18:30:27.093734 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-nvnm5_5536b1ca-ecee-4f69-9884-539b000d00f0/controller/0.log" Dec 01 18:30:27 crc kubenswrapper[4868]: I1201 18:30:27.293562 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-frr-files/0.log" Dec 01 18:30:27 crc kubenswrapper[4868]: I1201 18:30:27.454725 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-frr-files/0.log" Dec 01 18:30:27 crc kubenswrapper[4868]: I1201 18:30:27.454844 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-reloader/0.log" Dec 01 18:30:27 crc kubenswrapper[4868]: I1201 18:30:27.455931 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-metrics/0.log" Dec 01 18:30:27 crc kubenswrapper[4868]: I1201 18:30:27.510818 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-reloader/0.log" Dec 01 18:30:27 crc kubenswrapper[4868]: I1201 18:30:27.849394 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-metrics/0.log" Dec 01 18:30:27 crc kubenswrapper[4868]: I1201 18:30:27.851798 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-reloader/0.log" Dec 01 18:30:27 crc kubenswrapper[4868]: I1201 18:30:27.919493 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-metrics/0.log" Dec 01 18:30:28 crc kubenswrapper[4868]: I1201 18:30:28.070638 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-frr-files/0.log" Dec 01 18:30:28 crc kubenswrapper[4868]: I1201 18:30:28.246764 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-frr-files/0.log" Dec 01 18:30:28 crc kubenswrapper[4868]: I1201 18:30:28.249221 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-metrics/0.log" Dec 01 18:30:28 crc kubenswrapper[4868]: I1201 18:30:28.273862 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/cp-reloader/0.log" Dec 01 18:30:28 crc kubenswrapper[4868]: I1201 18:30:28.303363 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/controller/0.log" Dec 01 18:30:28 crc kubenswrapper[4868]: I1201 18:30:28.413758 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/frr-metrics/0.log" Dec 01 18:30:28 crc kubenswrapper[4868]: I1201 18:30:28.464584 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/kube-rbac-proxy/0.log" Dec 01 18:30:28 crc kubenswrapper[4868]: I1201 18:30:28.552181 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/kube-rbac-proxy-frr/0.log" Dec 01 18:30:28 crc kubenswrapper[4868]: I1201 18:30:28.669803 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/reloader/0.log" Dec 01 18:30:28 crc kubenswrapper[4868]: I1201 18:30:28.746662 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-54t6t_e22c99ca-b516-4e36-bc9b-ba5355170fd8/frr-k8s-webhook-server/0.log" Dec 01 18:30:28 crc kubenswrapper[4868]: I1201 18:30:28.896776 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-779b644dbb-f6925_0724b3ed-ae00-455d-9f48-94cd89bf56d9/manager/0.log" Dec 01 18:30:29 crc kubenswrapper[4868]: I1201 18:30:29.152396 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-85f6846bc9-2b4v6_f0ad7ab5-3180-42c7-90a1-7a75fac7a61d/webhook-server/0.log" Dec 01 18:30:29 crc kubenswrapper[4868]: I1201 18:30:29.296078 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xfljl_22663c8b-3ec6-453c-bb25-ca8d92576d07/kube-rbac-proxy/0.log" Dec 01 18:30:29 crc kubenswrapper[4868]: I1201 18:30:29.466238 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-57x2t_86bc0b53-b3fa-4c6d-b28b-d9ddbf9d1b41/frr/0.log" Dec 01 18:30:30 crc kubenswrapper[4868]: I1201 18:30:30.137440 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xfljl_22663c8b-3ec6-453c-bb25-ca8d92576d07/speaker/0.log" Dec 01 18:30:37 crc kubenswrapper[4868]: I1201 18:30:37.172659 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:30:37 crc kubenswrapper[4868]: E1201 18:30:37.173388 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:30:41 crc kubenswrapper[4868]: I1201 18:30:41.289397 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6_e1d4847d-1891-4e4c-9d68-04181da92d62/util/0.log" Dec 01 18:30:41 crc kubenswrapper[4868]: I1201 18:30:41.547326 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6_e1d4847d-1891-4e4c-9d68-04181da92d62/pull/0.log" Dec 01 18:30:41 crc kubenswrapper[4868]: I1201 18:30:41.549723 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6_e1d4847d-1891-4e4c-9d68-04181da92d62/pull/0.log" Dec 01 18:30:41 crc kubenswrapper[4868]: I1201 18:30:41.557308 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6_e1d4847d-1891-4e4c-9d68-04181da92d62/util/0.log" Dec 01 18:30:41 crc kubenswrapper[4868]: I1201 18:30:41.735191 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6_e1d4847d-1891-4e4c-9d68-04181da92d62/extract/0.log" Dec 01 18:30:41 crc kubenswrapper[4868]: I1201 18:30:41.736158 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6_e1d4847d-1891-4e4c-9d68-04181da92d62/util/0.log" Dec 01 18:30:41 crc kubenswrapper[4868]: I1201 18:30:41.762241 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fjvqh6_e1d4847d-1891-4e4c-9d68-04181da92d62/pull/0.log" Dec 01 18:30:41 crc kubenswrapper[4868]: I1201 18:30:41.889575 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv_22daded3-8016-41cc-bc8b-0dfc2ac3c093/util/0.log" Dec 01 18:30:42 crc kubenswrapper[4868]: I1201 18:30:42.048619 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv_22daded3-8016-41cc-bc8b-0dfc2ac3c093/pull/0.log" Dec 01 18:30:42 crc kubenswrapper[4868]: I1201 18:30:42.048826 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv_22daded3-8016-41cc-bc8b-0dfc2ac3c093/util/0.log" Dec 01 18:30:42 crc kubenswrapper[4868]: I1201 18:30:42.081559 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv_22daded3-8016-41cc-bc8b-0dfc2ac3c093/pull/0.log" Dec 01 18:30:42 crc kubenswrapper[4868]: I1201 18:30:42.254284 4868 scope.go:117] "RemoveContainer" containerID="9e450d322b643e6d9bcb8713507f4ad95b98cbf83227044c5a06618af0adff69" Dec 01 18:30:42 crc kubenswrapper[4868]: I1201 18:30:42.316197 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv_22daded3-8016-41cc-bc8b-0dfc2ac3c093/extract/0.log" Dec 01 18:30:42 crc kubenswrapper[4868]: I1201 18:30:42.332146 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv_22daded3-8016-41cc-bc8b-0dfc2ac3c093/pull/0.log" Dec 01 18:30:42 crc kubenswrapper[4868]: I1201 18:30:42.380414 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f838nlcv_22daded3-8016-41cc-bc8b-0dfc2ac3c093/util/0.log" Dec 01 18:30:42 crc kubenswrapper[4868]: I1201 18:30:42.488135 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-znxj6_2faafa31-ca8e-4da4-a1b4-e3d2709b4648/extract-utilities/0.log" Dec 01 18:30:42 crc kubenswrapper[4868]: I1201 18:30:42.668004 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-znxj6_2faafa31-ca8e-4da4-a1b4-e3d2709b4648/extract-utilities/0.log" Dec 01 18:30:42 crc kubenswrapper[4868]: I1201 18:30:42.722634 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-znxj6_2faafa31-ca8e-4da4-a1b4-e3d2709b4648/extract-content/0.log" Dec 01 18:30:42 crc kubenswrapper[4868]: I1201 18:30:42.742756 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-znxj6_2faafa31-ca8e-4da4-a1b4-e3d2709b4648/extract-content/0.log" Dec 01 18:30:42 crc kubenswrapper[4868]: I1201 18:30:42.884931 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-znxj6_2faafa31-ca8e-4da4-a1b4-e3d2709b4648/extract-content/0.log" Dec 01 18:30:42 crc kubenswrapper[4868]: I1201 18:30:42.928465 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-znxj6_2faafa31-ca8e-4da4-a1b4-e3d2709b4648/extract-utilities/0.log" Dec 01 18:30:43 crc kubenswrapper[4868]: I1201 18:30:43.154235 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fpxc7_ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a/extract-utilities/0.log" Dec 01 18:30:43 crc kubenswrapper[4868]: I1201 18:30:43.280396 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fpxc7_ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a/extract-utilities/0.log" Dec 01 18:30:43 crc kubenswrapper[4868]: I1201 18:30:43.304443 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fpxc7_ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a/extract-content/0.log" Dec 01 18:30:43 crc kubenswrapper[4868]: I1201 18:30:43.413468 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fpxc7_ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a/extract-content/0.log" Dec 01 18:30:43 crc kubenswrapper[4868]: I1201 18:30:43.470327 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-znxj6_2faafa31-ca8e-4da4-a1b4-e3d2709b4648/registry-server/0.log" Dec 01 18:30:43 crc kubenswrapper[4868]: I1201 18:30:43.584809 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fpxc7_ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a/extract-utilities/0.log" Dec 01 18:30:43 crc kubenswrapper[4868]: I1201 18:30:43.610615 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fpxc7_ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a/extract-content/0.log" Dec 01 18:30:43 crc kubenswrapper[4868]: I1201 18:30:43.779869 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4nfp5_0c043a8c-1af1-4aa0-99f9-0f6ebe27238a/marketplace-operator/0.log" Dec 01 18:30:43 crc kubenswrapper[4868]: I1201 18:30:43.825840 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fpxc7_ef5d10f1-ffa4-4554-83ab-6cf478e3ce4a/registry-server/0.log" Dec 01 18:30:43 crc kubenswrapper[4868]: I1201 18:30:43.970963 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9kf2c_5bff6f35-58ce-48e1-ae66-e29973c57a3f/extract-utilities/0.log" Dec 01 18:30:44 crc kubenswrapper[4868]: I1201 18:30:44.105007 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9kf2c_5bff6f35-58ce-48e1-ae66-e29973c57a3f/extract-content/0.log" Dec 01 18:30:44 crc kubenswrapper[4868]: I1201 18:30:44.108089 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9kf2c_5bff6f35-58ce-48e1-ae66-e29973c57a3f/extract-utilities/0.log" Dec 01 18:30:44 crc kubenswrapper[4868]: I1201 18:30:44.129323 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9kf2c_5bff6f35-58ce-48e1-ae66-e29973c57a3f/extract-content/0.log" Dec 01 18:30:44 crc kubenswrapper[4868]: I1201 18:30:44.326161 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9kf2c_5bff6f35-58ce-48e1-ae66-e29973c57a3f/extract-utilities/0.log" Dec 01 18:30:44 crc kubenswrapper[4868]: I1201 18:30:44.326975 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9kf2c_5bff6f35-58ce-48e1-ae66-e29973c57a3f/extract-content/0.log" Dec 01 18:30:44 crc kubenswrapper[4868]: I1201 18:30:44.465381 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9kf2c_5bff6f35-58ce-48e1-ae66-e29973c57a3f/registry-server/0.log" Dec 01 18:30:44 crc kubenswrapper[4868]: I1201 18:30:44.498537 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rcxdx_a904bedc-9d01-4a00-82c1-6f2826a33d81/extract-utilities/0.log" Dec 01 18:30:44 crc kubenswrapper[4868]: I1201 18:30:44.679045 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rcxdx_a904bedc-9d01-4a00-82c1-6f2826a33d81/extract-utilities/0.log" Dec 01 18:30:44 crc kubenswrapper[4868]: I1201 18:30:44.695232 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rcxdx_a904bedc-9d01-4a00-82c1-6f2826a33d81/extract-content/0.log" Dec 01 18:30:44 crc kubenswrapper[4868]: I1201 18:30:44.711431 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rcxdx_a904bedc-9d01-4a00-82c1-6f2826a33d81/extract-content/0.log" Dec 01 18:30:44 crc kubenswrapper[4868]: I1201 18:30:44.871491 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rcxdx_a904bedc-9d01-4a00-82c1-6f2826a33d81/extract-utilities/0.log" Dec 01 18:30:44 crc kubenswrapper[4868]: I1201 18:30:44.893668 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rcxdx_a904bedc-9d01-4a00-82c1-6f2826a33d81/extract-content/0.log" Dec 01 18:30:45 crc kubenswrapper[4868]: I1201 18:30:45.156648 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rcxdx_a904bedc-9d01-4a00-82c1-6f2826a33d81/registry-server/0.log" Dec 01 18:30:48 crc kubenswrapper[4868]: I1201 18:30:48.180970 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:30:48 crc kubenswrapper[4868]: E1201 18:30:48.181547 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:31:03 crc kubenswrapper[4868]: I1201 18:31:03.171586 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:31:03 crc kubenswrapper[4868]: E1201 18:31:03.172388 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:31:15 crc kubenswrapper[4868]: I1201 18:31:15.172143 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:31:15 crc kubenswrapper[4868]: E1201 18:31:15.173719 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:31:27 crc kubenswrapper[4868]: I1201 18:31:27.172844 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:31:27 crc kubenswrapper[4868]: E1201 18:31:27.173375 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:31:39 crc kubenswrapper[4868]: I1201 18:31:39.171474 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:31:39 crc kubenswrapper[4868]: E1201 18:31:39.172120 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:31:52 crc kubenswrapper[4868]: I1201 18:31:52.172695 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:31:52 crc kubenswrapper[4868]: E1201 18:31:52.173513 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:32:04 crc kubenswrapper[4868]: I1201 18:32:04.191147 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:32:04 crc kubenswrapper[4868]: E1201 18:32:04.198839 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:32:16 crc kubenswrapper[4868]: I1201 18:32:16.172438 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:32:16 crc kubenswrapper[4868]: E1201 18:32:16.173184 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:32:25 crc kubenswrapper[4868]: I1201 18:32:25.355345 4868 generic.go:334] "Generic (PLEG): container finished" podID="1ef4472c-7831-4220-a58e-efb68fa8d78e" containerID="23af67d3eeccc958f36af210b7e3028db3782016d453cde6766523faaddb467e" exitCode=0 Dec 01 18:32:25 crc kubenswrapper[4868]: I1201 18:32:25.355490 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vddj9/must-gather-k8hct" event={"ID":"1ef4472c-7831-4220-a58e-efb68fa8d78e","Type":"ContainerDied","Data":"23af67d3eeccc958f36af210b7e3028db3782016d453cde6766523faaddb467e"} Dec 01 18:32:25 crc kubenswrapper[4868]: I1201 18:32:25.356366 4868 scope.go:117] "RemoveContainer" containerID="23af67d3eeccc958f36af210b7e3028db3782016d453cde6766523faaddb467e" Dec 01 18:32:25 crc kubenswrapper[4868]: I1201 18:32:25.445764 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vddj9_must-gather-k8hct_1ef4472c-7831-4220-a58e-efb68fa8d78e/gather/0.log" Dec 01 18:32:28 crc kubenswrapper[4868]: I1201 18:32:28.182031 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:32:28 crc kubenswrapper[4868]: E1201 18:32:28.182708 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:32:35 crc kubenswrapper[4868]: I1201 18:32:35.843832 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vddj9/must-gather-k8hct"] Dec 01 18:32:35 crc kubenswrapper[4868]: I1201 18:32:35.844790 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-vddj9/must-gather-k8hct" podUID="1ef4472c-7831-4220-a58e-efb68fa8d78e" containerName="copy" containerID="cri-o://7b5a03b0af4f6f9fe18b4ff51a9da770ebf19f2e548907c2839ee4f51b62ffb0" gracePeriod=2 Dec 01 18:32:35 crc kubenswrapper[4868]: I1201 18:32:35.858619 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vddj9/must-gather-k8hct"] Dec 01 18:32:35 crc kubenswrapper[4868]: I1201 18:32:35.940154 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qnscj"] Dec 01 18:32:35 crc kubenswrapper[4868]: E1201 18:32:35.940507 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ef4472c-7831-4220-a58e-efb68fa8d78e" containerName="copy" Dec 01 18:32:35 crc kubenswrapper[4868]: I1201 18:32:35.940524 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ef4472c-7831-4220-a58e-efb68fa8d78e" containerName="copy" Dec 01 18:32:35 crc kubenswrapper[4868]: E1201 18:32:35.940538 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ef4472c-7831-4220-a58e-efb68fa8d78e" containerName="gather" Dec 01 18:32:35 crc kubenswrapper[4868]: I1201 18:32:35.940545 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ef4472c-7831-4220-a58e-efb68fa8d78e" containerName="gather" Dec 01 18:32:35 crc kubenswrapper[4868]: E1201 18:32:35.941656 4868 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="099c1666-e6f3-4192-b083-b2400e6e6c68" containerName="collect-profiles" Dec 01 18:32:35 crc kubenswrapper[4868]: I1201 18:32:35.941675 4868 state_mem.go:107] "Deleted CPUSet assignment" podUID="099c1666-e6f3-4192-b083-b2400e6e6c68" containerName="collect-profiles" Dec 01 18:32:35 crc kubenswrapper[4868]: I1201 18:32:35.941840 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ef4472c-7831-4220-a58e-efb68fa8d78e" containerName="gather" Dec 01 18:32:35 crc kubenswrapper[4868]: I1201 18:32:35.941859 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="099c1666-e6f3-4192-b083-b2400e6e6c68" containerName="collect-profiles" Dec 01 18:32:35 crc kubenswrapper[4868]: I1201 18:32:35.941874 4868 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ef4472c-7831-4220-a58e-efb68fa8d78e" containerName="copy" Dec 01 18:32:35 crc kubenswrapper[4868]: I1201 18:32:35.943706 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:35 crc kubenswrapper[4868]: I1201 18:32:35.955059 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qnscj"] Dec 01 18:32:35 crc kubenswrapper[4868]: I1201 18:32:35.982062 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d910d7-6569-4473-99fe-0d1efe7c618f-catalog-content\") pod \"redhat-operators-qnscj\" (UID: \"24d910d7-6569-4473-99fe-0d1efe7c618f\") " pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:35 crc kubenswrapper[4868]: I1201 18:32:35.982238 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvqls\" (UniqueName: \"kubernetes.io/projected/24d910d7-6569-4473-99fe-0d1efe7c618f-kube-api-access-wvqls\") pod \"redhat-operators-qnscj\" (UID: \"24d910d7-6569-4473-99fe-0d1efe7c618f\") " pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:35 crc kubenswrapper[4868]: I1201 18:32:35.982315 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d910d7-6569-4473-99fe-0d1efe7c618f-utilities\") pod \"redhat-operators-qnscj\" (UID: \"24d910d7-6569-4473-99fe-0d1efe7c618f\") " pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.084721 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvqls\" (UniqueName: \"kubernetes.io/projected/24d910d7-6569-4473-99fe-0d1efe7c618f-kube-api-access-wvqls\") pod \"redhat-operators-qnscj\" (UID: \"24d910d7-6569-4473-99fe-0d1efe7c618f\") " pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.084778 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d910d7-6569-4473-99fe-0d1efe7c618f-utilities\") pod \"redhat-operators-qnscj\" (UID: \"24d910d7-6569-4473-99fe-0d1efe7c618f\") " pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.084880 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d910d7-6569-4473-99fe-0d1efe7c618f-catalog-content\") pod \"redhat-operators-qnscj\" (UID: \"24d910d7-6569-4473-99fe-0d1efe7c618f\") " pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.085330 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d910d7-6569-4473-99fe-0d1efe7c618f-catalog-content\") pod \"redhat-operators-qnscj\" (UID: \"24d910d7-6569-4473-99fe-0d1efe7c618f\") " pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.085807 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d910d7-6569-4473-99fe-0d1efe7c618f-utilities\") pod \"redhat-operators-qnscj\" (UID: \"24d910d7-6569-4473-99fe-0d1efe7c618f\") " pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.115421 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvqls\" (UniqueName: \"kubernetes.io/projected/24d910d7-6569-4473-99fe-0d1efe7c618f-kube-api-access-wvqls\") pod \"redhat-operators-qnscj\" (UID: \"24d910d7-6569-4473-99fe-0d1efe7c618f\") " pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.314391 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.403400 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vddj9_must-gather-k8hct_1ef4472c-7831-4220-a58e-efb68fa8d78e/copy/0.log" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.403908 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/must-gather-k8hct" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.491884 4868 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vddj9_must-gather-k8hct_1ef4472c-7831-4220-a58e-efb68fa8d78e/copy/0.log" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.492603 4868 generic.go:334] "Generic (PLEG): container finished" podID="1ef4472c-7831-4220-a58e-efb68fa8d78e" containerID="7b5a03b0af4f6f9fe18b4ff51a9da770ebf19f2e548907c2839ee4f51b62ffb0" exitCode=143 Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.492658 4868 scope.go:117] "RemoveContainer" containerID="7b5a03b0af4f6f9fe18b4ff51a9da770ebf19f2e548907c2839ee4f51b62ffb0" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.492855 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vddj9/must-gather-k8hct" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.523032 4868 scope.go:117] "RemoveContainer" containerID="23af67d3eeccc958f36af210b7e3028db3782016d453cde6766523faaddb467e" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.599152 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmckq\" (UniqueName: \"kubernetes.io/projected/1ef4472c-7831-4220-a58e-efb68fa8d78e-kube-api-access-fmckq\") pod \"1ef4472c-7831-4220-a58e-efb68fa8d78e\" (UID: \"1ef4472c-7831-4220-a58e-efb68fa8d78e\") " Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.599537 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1ef4472c-7831-4220-a58e-efb68fa8d78e-must-gather-output\") pod \"1ef4472c-7831-4220-a58e-efb68fa8d78e\" (UID: \"1ef4472c-7831-4220-a58e-efb68fa8d78e\") " Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.608916 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ef4472c-7831-4220-a58e-efb68fa8d78e-kube-api-access-fmckq" (OuterVolumeSpecName: "kube-api-access-fmckq") pod "1ef4472c-7831-4220-a58e-efb68fa8d78e" (UID: "1ef4472c-7831-4220-a58e-efb68fa8d78e"). InnerVolumeSpecName "kube-api-access-fmckq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.646375 4868 scope.go:117] "RemoveContainer" containerID="7b5a03b0af4f6f9fe18b4ff51a9da770ebf19f2e548907c2839ee4f51b62ffb0" Dec 01 18:32:36 crc kubenswrapper[4868]: E1201 18:32:36.646813 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b5a03b0af4f6f9fe18b4ff51a9da770ebf19f2e548907c2839ee4f51b62ffb0\": container with ID starting with 7b5a03b0af4f6f9fe18b4ff51a9da770ebf19f2e548907c2839ee4f51b62ffb0 not found: ID does not exist" containerID="7b5a03b0af4f6f9fe18b4ff51a9da770ebf19f2e548907c2839ee4f51b62ffb0" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.646865 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b5a03b0af4f6f9fe18b4ff51a9da770ebf19f2e548907c2839ee4f51b62ffb0"} err="failed to get container status \"7b5a03b0af4f6f9fe18b4ff51a9da770ebf19f2e548907c2839ee4f51b62ffb0\": rpc error: code = NotFound desc = could not find container \"7b5a03b0af4f6f9fe18b4ff51a9da770ebf19f2e548907c2839ee4f51b62ffb0\": container with ID starting with 7b5a03b0af4f6f9fe18b4ff51a9da770ebf19f2e548907c2839ee4f51b62ffb0 not found: ID does not exist" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.646896 4868 scope.go:117] "RemoveContainer" containerID="23af67d3eeccc958f36af210b7e3028db3782016d453cde6766523faaddb467e" Dec 01 18:32:36 crc kubenswrapper[4868]: E1201 18:32:36.647291 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23af67d3eeccc958f36af210b7e3028db3782016d453cde6766523faaddb467e\": container with ID starting with 23af67d3eeccc958f36af210b7e3028db3782016d453cde6766523faaddb467e not found: ID does not exist" containerID="23af67d3eeccc958f36af210b7e3028db3782016d453cde6766523faaddb467e" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.647329 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23af67d3eeccc958f36af210b7e3028db3782016d453cde6766523faaddb467e"} err="failed to get container status \"23af67d3eeccc958f36af210b7e3028db3782016d453cde6766523faaddb467e\": rpc error: code = NotFound desc = could not find container \"23af67d3eeccc958f36af210b7e3028db3782016d453cde6766523faaddb467e\": container with ID starting with 23af67d3eeccc958f36af210b7e3028db3782016d453cde6766523faaddb467e not found: ID does not exist" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.702446 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmckq\" (UniqueName: \"kubernetes.io/projected/1ef4472c-7831-4220-a58e-efb68fa8d78e-kube-api-access-fmckq\") on node \"crc\" DevicePath \"\"" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.770312 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ef4472c-7831-4220-a58e-efb68fa8d78e-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "1ef4472c-7831-4220-a58e-efb68fa8d78e" (UID: "1ef4472c-7831-4220-a58e-efb68fa8d78e"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.791443 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qnscj"] Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.804647 4868 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1ef4472c-7831-4220-a58e-efb68fa8d78e-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.935274 4868 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-blf2m"] Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.938761 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blf2m" Dec 01 18:32:36 crc kubenswrapper[4868]: I1201 18:32:36.952617 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-blf2m"] Dec 01 18:32:37 crc kubenswrapper[4868]: I1201 18:32:37.109685 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33334217-460a-48bc-8a0d-64d4143981b8-catalog-content\") pod \"certified-operators-blf2m\" (UID: \"33334217-460a-48bc-8a0d-64d4143981b8\") " pod="openshift-marketplace/certified-operators-blf2m" Dec 01 18:32:37 crc kubenswrapper[4868]: I1201 18:32:37.109765 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txg2j\" (UniqueName: \"kubernetes.io/projected/33334217-460a-48bc-8a0d-64d4143981b8-kube-api-access-txg2j\") pod \"certified-operators-blf2m\" (UID: \"33334217-460a-48bc-8a0d-64d4143981b8\") " pod="openshift-marketplace/certified-operators-blf2m" Dec 01 18:32:37 crc kubenswrapper[4868]: I1201 18:32:37.109822 4868 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33334217-460a-48bc-8a0d-64d4143981b8-utilities\") pod \"certified-operators-blf2m\" (UID: \"33334217-460a-48bc-8a0d-64d4143981b8\") " pod="openshift-marketplace/certified-operators-blf2m" Dec 01 18:32:37 crc kubenswrapper[4868]: I1201 18:32:37.212030 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33334217-460a-48bc-8a0d-64d4143981b8-catalog-content\") pod \"certified-operators-blf2m\" (UID: \"33334217-460a-48bc-8a0d-64d4143981b8\") " pod="openshift-marketplace/certified-operators-blf2m" Dec 01 18:32:37 crc kubenswrapper[4868]: I1201 18:32:37.212134 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txg2j\" (UniqueName: \"kubernetes.io/projected/33334217-460a-48bc-8a0d-64d4143981b8-kube-api-access-txg2j\") pod \"certified-operators-blf2m\" (UID: \"33334217-460a-48bc-8a0d-64d4143981b8\") " pod="openshift-marketplace/certified-operators-blf2m" Dec 01 18:32:37 crc kubenswrapper[4868]: I1201 18:32:37.212191 4868 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33334217-460a-48bc-8a0d-64d4143981b8-utilities\") pod \"certified-operators-blf2m\" (UID: \"33334217-460a-48bc-8a0d-64d4143981b8\") " pod="openshift-marketplace/certified-operators-blf2m" Dec 01 18:32:37 crc kubenswrapper[4868]: I1201 18:32:37.212662 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33334217-460a-48bc-8a0d-64d4143981b8-catalog-content\") pod \"certified-operators-blf2m\" (UID: \"33334217-460a-48bc-8a0d-64d4143981b8\") " pod="openshift-marketplace/certified-operators-blf2m" Dec 01 18:32:37 crc kubenswrapper[4868]: I1201 18:32:37.212711 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33334217-460a-48bc-8a0d-64d4143981b8-utilities\") pod \"certified-operators-blf2m\" (UID: \"33334217-460a-48bc-8a0d-64d4143981b8\") " pod="openshift-marketplace/certified-operators-blf2m" Dec 01 18:32:37 crc kubenswrapper[4868]: I1201 18:32:37.231677 4868 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txg2j\" (UniqueName: \"kubernetes.io/projected/33334217-460a-48bc-8a0d-64d4143981b8-kube-api-access-txg2j\") pod \"certified-operators-blf2m\" (UID: \"33334217-460a-48bc-8a0d-64d4143981b8\") " pod="openshift-marketplace/certified-operators-blf2m" Dec 01 18:32:37 crc kubenswrapper[4868]: I1201 18:32:37.391861 4868 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-blf2m" Dec 01 18:32:37 crc kubenswrapper[4868]: I1201 18:32:37.502224 4868 generic.go:334] "Generic (PLEG): container finished" podID="24d910d7-6569-4473-99fe-0d1efe7c618f" containerID="a7b841b74c826a932537af25dc069273582289e78f507a8a888611bdaff83a82" exitCode=0 Dec 01 18:32:37 crc kubenswrapper[4868]: I1201 18:32:37.502289 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnscj" event={"ID":"24d910d7-6569-4473-99fe-0d1efe7c618f","Type":"ContainerDied","Data":"a7b841b74c826a932537af25dc069273582289e78f507a8a888611bdaff83a82"} Dec 01 18:32:37 crc kubenswrapper[4868]: I1201 18:32:37.502317 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnscj" event={"ID":"24d910d7-6569-4473-99fe-0d1efe7c618f","Type":"ContainerStarted","Data":"78edaf0c1eca1f89484210010d4d6a4eb99327961f6856f2594109c75436d8d2"} Dec 01 18:32:38 crc kubenswrapper[4868]: I1201 18:32:38.049397 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-blf2m"] Dec 01 18:32:38 crc kubenswrapper[4868]: W1201 18:32:38.056038 4868 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33334217_460a_48bc_8a0d_64d4143981b8.slice/crio-37c5ecf45d894ecc5e03825760fe3c931bf0ffa96618d71790d55c0886e17b91 WatchSource:0}: Error finding container 37c5ecf45d894ecc5e03825760fe3c931bf0ffa96618d71790d55c0886e17b91: Status 404 returned error can't find the container with id 37c5ecf45d894ecc5e03825760fe3c931bf0ffa96618d71790d55c0886e17b91 Dec 01 18:32:38 crc kubenswrapper[4868]: I1201 18:32:38.189205 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ef4472c-7831-4220-a58e-efb68fa8d78e" path="/var/lib/kubelet/pods/1ef4472c-7831-4220-a58e-efb68fa8d78e/volumes" Dec 01 18:32:38 crc kubenswrapper[4868]: I1201 18:32:38.519360 4868 generic.go:334] "Generic (PLEG): container finished" podID="33334217-460a-48bc-8a0d-64d4143981b8" containerID="43b24696a2716fcf3585b688de142e18a520047eafd19d4c10787a2bedb20025" exitCode=0 Dec 01 18:32:38 crc kubenswrapper[4868]: I1201 18:32:38.519414 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blf2m" event={"ID":"33334217-460a-48bc-8a0d-64d4143981b8","Type":"ContainerDied","Data":"43b24696a2716fcf3585b688de142e18a520047eafd19d4c10787a2bedb20025"} Dec 01 18:32:38 crc kubenswrapper[4868]: I1201 18:32:38.519448 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blf2m" event={"ID":"33334217-460a-48bc-8a0d-64d4143981b8","Type":"ContainerStarted","Data":"37c5ecf45d894ecc5e03825760fe3c931bf0ffa96618d71790d55c0886e17b91"} Dec 01 18:32:39 crc kubenswrapper[4868]: I1201 18:32:39.537147 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnscj" event={"ID":"24d910d7-6569-4473-99fe-0d1efe7c618f","Type":"ContainerStarted","Data":"75e32f2c8ee8893891fdd782c542159ce988b4685264fd954dbf6861b563f0b7"} Dec 01 18:32:40 crc kubenswrapper[4868]: I1201 18:32:40.548075 4868 generic.go:334] "Generic (PLEG): container finished" podID="24d910d7-6569-4473-99fe-0d1efe7c618f" containerID="75e32f2c8ee8893891fdd782c542159ce988b4685264fd954dbf6861b563f0b7" exitCode=0 Dec 01 18:32:40 crc kubenswrapper[4868]: I1201 18:32:40.548122 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnscj" event={"ID":"24d910d7-6569-4473-99fe-0d1efe7c618f","Type":"ContainerDied","Data":"75e32f2c8ee8893891fdd782c542159ce988b4685264fd954dbf6861b563f0b7"} Dec 01 18:32:40 crc kubenswrapper[4868]: I1201 18:32:40.550414 4868 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 18:32:41 crc kubenswrapper[4868]: I1201 18:32:41.173309 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:32:41 crc kubenswrapper[4868]: E1201 18:32:41.174189 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:32:43 crc kubenswrapper[4868]: I1201 18:32:43.576493 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnscj" event={"ID":"24d910d7-6569-4473-99fe-0d1efe7c618f","Type":"ContainerStarted","Data":"95a7c1a96f51eb6537e1e17a1f7f200ef647bf7037ea3779dc685257ff533017"} Dec 01 18:32:43 crc kubenswrapper[4868]: I1201 18:32:43.601656 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qnscj" podStartSLOduration=5.109302777 podStartE2EDuration="8.601632847s" podCreationTimestamp="2025-12-01 18:32:35 +0000 UTC" firstStartedPulling="2025-12-01 18:32:37.504576919 +0000 UTC m=+4029.875687330" lastFinishedPulling="2025-12-01 18:32:40.996906979 +0000 UTC m=+4033.368017400" observedRunningTime="2025-12-01 18:32:43.596380146 +0000 UTC m=+4035.967490577" watchObservedRunningTime="2025-12-01 18:32:43.601632847 +0000 UTC m=+4035.972743248" Dec 01 18:32:44 crc kubenswrapper[4868]: I1201 18:32:44.586137 4868 generic.go:334] "Generic (PLEG): container finished" podID="33334217-460a-48bc-8a0d-64d4143981b8" containerID="77532f63631ee901eec25af32a9da5fee00319a58ee742165f843110d74403bc" exitCode=0 Dec 01 18:32:44 crc kubenswrapper[4868]: I1201 18:32:44.586279 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blf2m" event={"ID":"33334217-460a-48bc-8a0d-64d4143981b8","Type":"ContainerDied","Data":"77532f63631ee901eec25af32a9da5fee00319a58ee742165f843110d74403bc"} Dec 01 18:32:46 crc kubenswrapper[4868]: I1201 18:32:46.315126 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:46 crc kubenswrapper[4868]: I1201 18:32:46.315471 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:47 crc kubenswrapper[4868]: I1201 18:32:47.373810 4868 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qnscj" podUID="24d910d7-6569-4473-99fe-0d1efe7c618f" containerName="registry-server" probeResult="failure" output=< Dec 01 18:32:47 crc kubenswrapper[4868]: timeout: failed to connect service ":50051" within 1s Dec 01 18:32:47 crc kubenswrapper[4868]: > Dec 01 18:32:47 crc kubenswrapper[4868]: I1201 18:32:47.619557 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-blf2m" event={"ID":"33334217-460a-48bc-8a0d-64d4143981b8","Type":"ContainerStarted","Data":"d17e758076785031ed3108b8989ee3c4894f17611c4db5c0b6da4acad30576c2"} Dec 01 18:32:47 crc kubenswrapper[4868]: I1201 18:32:47.648662 4868 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-blf2m" podStartSLOduration=3.654215992 podStartE2EDuration="11.648632685s" podCreationTimestamp="2025-12-01 18:32:36 +0000 UTC" firstStartedPulling="2025-12-01 18:32:38.521176277 +0000 UTC m=+4030.892286688" lastFinishedPulling="2025-12-01 18:32:46.51559297 +0000 UTC m=+4038.886703381" observedRunningTime="2025-12-01 18:32:47.637836413 +0000 UTC m=+4040.008946824" watchObservedRunningTime="2025-12-01 18:32:47.648632685 +0000 UTC m=+4040.019743096" Dec 01 18:32:52 crc kubenswrapper[4868]: I1201 18:32:52.172507 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:32:52 crc kubenswrapper[4868]: E1201 18:32:52.174077 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:32:56 crc kubenswrapper[4868]: I1201 18:32:56.358217 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:56 crc kubenswrapper[4868]: I1201 18:32:56.411730 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:56 crc kubenswrapper[4868]: I1201 18:32:56.593112 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qnscj"] Dec 01 18:32:57 crc kubenswrapper[4868]: I1201 18:32:57.392244 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-blf2m" Dec 01 18:32:57 crc kubenswrapper[4868]: I1201 18:32:57.392285 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-blf2m" Dec 01 18:32:57 crc kubenswrapper[4868]: I1201 18:32:57.444108 4868 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-blf2m" Dec 01 18:32:57 crc kubenswrapper[4868]: I1201 18:32:57.739767 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qnscj" podUID="24d910d7-6569-4473-99fe-0d1efe7c618f" containerName="registry-server" containerID="cri-o://95a7c1a96f51eb6537e1e17a1f7f200ef647bf7037ea3779dc685257ff533017" gracePeriod=2 Dec 01 18:32:57 crc kubenswrapper[4868]: I1201 18:32:57.796258 4868 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-blf2m" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.203282 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.364890 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d910d7-6569-4473-99fe-0d1efe7c618f-catalog-content\") pod \"24d910d7-6569-4473-99fe-0d1efe7c618f\" (UID: \"24d910d7-6569-4473-99fe-0d1efe7c618f\") " Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.365443 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d910d7-6569-4473-99fe-0d1efe7c618f-utilities\") pod \"24d910d7-6569-4473-99fe-0d1efe7c618f\" (UID: \"24d910d7-6569-4473-99fe-0d1efe7c618f\") " Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.365556 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvqls\" (UniqueName: \"kubernetes.io/projected/24d910d7-6569-4473-99fe-0d1efe7c618f-kube-api-access-wvqls\") pod \"24d910d7-6569-4473-99fe-0d1efe7c618f\" (UID: \"24d910d7-6569-4473-99fe-0d1efe7c618f\") " Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.366056 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24d910d7-6569-4473-99fe-0d1efe7c618f-utilities" (OuterVolumeSpecName: "utilities") pod "24d910d7-6569-4473-99fe-0d1efe7c618f" (UID: "24d910d7-6569-4473-99fe-0d1efe7c618f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.366142 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d910d7-6569-4473-99fe-0d1efe7c618f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.370283 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24d910d7-6569-4473-99fe-0d1efe7c618f-kube-api-access-wvqls" (OuterVolumeSpecName: "kube-api-access-wvqls") pod "24d910d7-6569-4473-99fe-0d1efe7c618f" (UID: "24d910d7-6569-4473-99fe-0d1efe7c618f"). InnerVolumeSpecName "kube-api-access-wvqls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.467651 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvqls\" (UniqueName: \"kubernetes.io/projected/24d910d7-6569-4473-99fe-0d1efe7c618f-kube-api-access-wvqls\") on node \"crc\" DevicePath \"\"" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.480986 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24d910d7-6569-4473-99fe-0d1efe7c618f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24d910d7-6569-4473-99fe-0d1efe7c618f" (UID: "24d910d7-6569-4473-99fe-0d1efe7c618f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.569362 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d910d7-6569-4473-99fe-0d1efe7c618f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.748551 4868 generic.go:334] "Generic (PLEG): container finished" podID="24d910d7-6569-4473-99fe-0d1efe7c618f" containerID="95a7c1a96f51eb6537e1e17a1f7f200ef647bf7037ea3779dc685257ff533017" exitCode=0 Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.748603 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qnscj" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.748642 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnscj" event={"ID":"24d910d7-6569-4473-99fe-0d1efe7c618f","Type":"ContainerDied","Data":"95a7c1a96f51eb6537e1e17a1f7f200ef647bf7037ea3779dc685257ff533017"} Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.748692 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnscj" event={"ID":"24d910d7-6569-4473-99fe-0d1efe7c618f","Type":"ContainerDied","Data":"78edaf0c1eca1f89484210010d4d6a4eb99327961f6856f2594109c75436d8d2"} Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.748715 4868 scope.go:117] "RemoveContainer" containerID="95a7c1a96f51eb6537e1e17a1f7f200ef647bf7037ea3779dc685257ff533017" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.765119 4868 scope.go:117] "RemoveContainer" containerID="75e32f2c8ee8893891fdd782c542159ce988b4685264fd954dbf6861b563f0b7" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.782031 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qnscj"] Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.793469 4868 scope.go:117] "RemoveContainer" containerID="a7b841b74c826a932537af25dc069273582289e78f507a8a888611bdaff83a82" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.811004 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qnscj"] Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.823780 4868 scope.go:117] "RemoveContainer" containerID="95a7c1a96f51eb6537e1e17a1f7f200ef647bf7037ea3779dc685257ff533017" Dec 01 18:32:58 crc kubenswrapper[4868]: E1201 18:32:58.824282 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95a7c1a96f51eb6537e1e17a1f7f200ef647bf7037ea3779dc685257ff533017\": container with ID starting with 95a7c1a96f51eb6537e1e17a1f7f200ef647bf7037ea3779dc685257ff533017 not found: ID does not exist" containerID="95a7c1a96f51eb6537e1e17a1f7f200ef647bf7037ea3779dc685257ff533017" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.824313 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95a7c1a96f51eb6537e1e17a1f7f200ef647bf7037ea3779dc685257ff533017"} err="failed to get container status \"95a7c1a96f51eb6537e1e17a1f7f200ef647bf7037ea3779dc685257ff533017\": rpc error: code = NotFound desc = could not find container \"95a7c1a96f51eb6537e1e17a1f7f200ef647bf7037ea3779dc685257ff533017\": container with ID starting with 95a7c1a96f51eb6537e1e17a1f7f200ef647bf7037ea3779dc685257ff533017 not found: ID does not exist" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.824334 4868 scope.go:117] "RemoveContainer" containerID="75e32f2c8ee8893891fdd782c542159ce988b4685264fd954dbf6861b563f0b7" Dec 01 18:32:58 crc kubenswrapper[4868]: E1201 18:32:58.824621 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75e32f2c8ee8893891fdd782c542159ce988b4685264fd954dbf6861b563f0b7\": container with ID starting with 75e32f2c8ee8893891fdd782c542159ce988b4685264fd954dbf6861b563f0b7 not found: ID does not exist" containerID="75e32f2c8ee8893891fdd782c542159ce988b4685264fd954dbf6861b563f0b7" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.824661 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75e32f2c8ee8893891fdd782c542159ce988b4685264fd954dbf6861b563f0b7"} err="failed to get container status \"75e32f2c8ee8893891fdd782c542159ce988b4685264fd954dbf6861b563f0b7\": rpc error: code = NotFound desc = could not find container \"75e32f2c8ee8893891fdd782c542159ce988b4685264fd954dbf6861b563f0b7\": container with ID starting with 75e32f2c8ee8893891fdd782c542159ce988b4685264fd954dbf6861b563f0b7 not found: ID does not exist" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.824688 4868 scope.go:117] "RemoveContainer" containerID="a7b841b74c826a932537af25dc069273582289e78f507a8a888611bdaff83a82" Dec 01 18:32:58 crc kubenswrapper[4868]: E1201 18:32:58.825110 4868 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7b841b74c826a932537af25dc069273582289e78f507a8a888611bdaff83a82\": container with ID starting with a7b841b74c826a932537af25dc069273582289e78f507a8a888611bdaff83a82 not found: ID does not exist" containerID="a7b841b74c826a932537af25dc069273582289e78f507a8a888611bdaff83a82" Dec 01 18:32:58 crc kubenswrapper[4868]: I1201 18:32:58.825133 4868 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7b841b74c826a932537af25dc069273582289e78f507a8a888611bdaff83a82"} err="failed to get container status \"a7b841b74c826a932537af25dc069273582289e78f507a8a888611bdaff83a82\": rpc error: code = NotFound desc = could not find container \"a7b841b74c826a932537af25dc069273582289e78f507a8a888611bdaff83a82\": container with ID starting with a7b841b74c826a932537af25dc069273582289e78f507a8a888611bdaff83a82 not found: ID does not exist" Dec 01 18:32:59 crc kubenswrapper[4868]: I1201 18:32:59.207546 4868 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-blf2m"] Dec 01 18:32:59 crc kubenswrapper[4868]: I1201 18:32:59.604398 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-znxj6"] Dec 01 18:32:59 crc kubenswrapper[4868]: I1201 18:32:59.604822 4868 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-znxj6" podUID="2faafa31-ca8e-4da4-a1b4-e3d2709b4648" containerName="registry-server" containerID="cri-o://6535c775409e77f58f8df53032a1d000c55894a88d6d0e340e911f758a67c8ca" gracePeriod=2 Dec 01 18:32:59 crc kubenswrapper[4868]: I1201 18:32:59.759363 4868 generic.go:334] "Generic (PLEG): container finished" podID="2faafa31-ca8e-4da4-a1b4-e3d2709b4648" containerID="6535c775409e77f58f8df53032a1d000c55894a88d6d0e340e911f758a67c8ca" exitCode=0 Dec 01 18:32:59 crc kubenswrapper[4868]: I1201 18:32:59.759464 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znxj6" event={"ID":"2faafa31-ca8e-4da4-a1b4-e3d2709b4648","Type":"ContainerDied","Data":"6535c775409e77f58f8df53032a1d000c55894a88d6d0e340e911f758a67c8ca"} Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.121499 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znxj6" Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.199682 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24d910d7-6569-4473-99fe-0d1efe7c618f" path="/var/lib/kubelet/pods/24d910d7-6569-4473-99fe-0d1efe7c618f/volumes" Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.206864 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwz6x\" (UniqueName: \"kubernetes.io/projected/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-kube-api-access-pwz6x\") pod \"2faafa31-ca8e-4da4-a1b4-e3d2709b4648\" (UID: \"2faafa31-ca8e-4da4-a1b4-e3d2709b4648\") " Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.207151 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-catalog-content\") pod \"2faafa31-ca8e-4da4-a1b4-e3d2709b4648\" (UID: \"2faafa31-ca8e-4da4-a1b4-e3d2709b4648\") " Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.207299 4868 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-utilities\") pod \"2faafa31-ca8e-4da4-a1b4-e3d2709b4648\" (UID: \"2faafa31-ca8e-4da4-a1b4-e3d2709b4648\") " Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.208390 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-utilities" (OuterVolumeSpecName: "utilities") pod "2faafa31-ca8e-4da4-a1b4-e3d2709b4648" (UID: "2faafa31-ca8e-4da4-a1b4-e3d2709b4648"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.216432 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-kube-api-access-pwz6x" (OuterVolumeSpecName: "kube-api-access-pwz6x") pod "2faafa31-ca8e-4da4-a1b4-e3d2709b4648" (UID: "2faafa31-ca8e-4da4-a1b4-e3d2709b4648"). InnerVolumeSpecName "kube-api-access-pwz6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.263932 4868 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2faafa31-ca8e-4da4-a1b4-e3d2709b4648" (UID: "2faafa31-ca8e-4da4-a1b4-e3d2709b4648"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.309895 4868 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.309930 4868 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.309957 4868 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwz6x\" (UniqueName: \"kubernetes.io/projected/2faafa31-ca8e-4da4-a1b4-e3d2709b4648-kube-api-access-pwz6x\") on node \"crc\" DevicePath \"\"" Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.771934 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znxj6" event={"ID":"2faafa31-ca8e-4da4-a1b4-e3d2709b4648","Type":"ContainerDied","Data":"1115711895dbf0833014baf2f5523dff2a5a1ccba8954e931b7a87d1fddfeec9"} Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.772020 4868 scope.go:117] "RemoveContainer" containerID="6535c775409e77f58f8df53032a1d000c55894a88d6d0e340e911f758a67c8ca" Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.772175 4868 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znxj6" Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.812258 4868 scope.go:117] "RemoveContainer" containerID="a013d6acaa5ac65f3bba8db55766eb4a784131fb28e4b1fc0da1e74b201da117" Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.812527 4868 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-znxj6"] Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.832459 4868 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-znxj6"] Dec 01 18:33:00 crc kubenswrapper[4868]: I1201 18:33:00.842638 4868 scope.go:117] "RemoveContainer" containerID="80079058bbb927e2310b6503faea9c0bccef225b615cbd38b8fdf5595d3b641d" Dec 01 18:33:02 crc kubenswrapper[4868]: I1201 18:33:02.183489 4868 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2faafa31-ca8e-4da4-a1b4-e3d2709b4648" path="/var/lib/kubelet/pods/2faafa31-ca8e-4da4-a1b4-e3d2709b4648/volumes" Dec 01 18:33:06 crc kubenswrapper[4868]: I1201 18:33:06.173398 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:33:06 crc kubenswrapper[4868]: E1201 18:33:06.174553 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:33:21 crc kubenswrapper[4868]: I1201 18:33:21.172394 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:33:21 crc kubenswrapper[4868]: E1201 18:33:21.173150 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:33:33 crc kubenswrapper[4868]: I1201 18:33:33.172601 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:33:33 crc kubenswrapper[4868]: E1201 18:33:33.173936 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:33:44 crc kubenswrapper[4868]: I1201 18:33:44.172888 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:33:44 crc kubenswrapper[4868]: E1201 18:33:44.174048 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:33:55 crc kubenswrapper[4868]: I1201 18:33:55.173003 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:33:55 crc kubenswrapper[4868]: E1201 18:33:55.174038 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:34:10 crc kubenswrapper[4868]: I1201 18:34:10.171967 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:34:10 crc kubenswrapper[4868]: E1201 18:34:10.172823 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:34:24 crc kubenswrapper[4868]: I1201 18:34:24.171997 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:34:24 crc kubenswrapper[4868]: E1201 18:34:24.172726 4868 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-njgpn_openshift-machine-config-operator(ec4e08bf-df56-4490-8e79-a4175c727405)\"" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" Dec 01 18:34:37 crc kubenswrapper[4868]: I1201 18:34:37.172505 4868 scope.go:117] "RemoveContainer" containerID="91de1412b80a1e0a0a3993ad07d79ec52427fa3899fac5c29b0b3c154e96148f" Dec 01 18:34:38 crc kubenswrapper[4868]: I1201 18:34:38.684373 4868 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" event={"ID":"ec4e08bf-df56-4490-8e79-a4175c727405","Type":"ContainerStarted","Data":"fc0eb905ffb76a59a562e8b4ec389c14b5013eae6baf141301eed914914eaf6d"} Dec 01 18:34:43 crc kubenswrapper[4868]: I1201 18:34:43.400992 4868 scope.go:117] "RemoveContainer" containerID="8660cae0d0a180faf55792f97e80a742d5829b6e1ce21c8d1aa59909e427f544" Dec 01 18:36:55 crc kubenswrapper[4868]: I1201 18:36:55.904364 4868 patch_prober.go:28] interesting pod/machine-config-daemon-njgpn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 18:36:55 crc kubenswrapper[4868]: I1201 18:36:55.905037 4868 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-njgpn" podUID="ec4e08bf-df56-4490-8e79-a4175c727405" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"